var/home/core/zuul-output/0000755000175000017500000000000015157205450014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015157214445015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000255104315157214366020272 0ustar corecoreikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gf ?"mv?_eGbuu񯷑7+%f?7ݭ7֫t% oo/q3m^]/o?8.7oW}ʋghewx/mX,ojŻ ^Tb3b#׳:}=p7뼝ca㑔`e0I1Q!&ѱ[/o^{W-{t3_U|6 x)K#/5ΌR"ggóisR)N %emOQ/Ϋ_oa0vs68/Jʢ ܚʂ9ss3+aô٥J}{37FEbп3 FKX1QRQlrTvb)E,s)Wɀ;$#LcdHM%vz_. o~I|3j dF{ "IΩ?PF~J~ ` 17ׅwڋًM)$Fiqw7Gt7L"u 0V9c  ˹dvYļU[ Z.׿/h QZ*U1|t5wKOؾ{mk b2 ܨ;RJK!b>JR*kl|+"N'C_#a7]d]sJg;;>Yp׫,w`ɚ'd$ecwŻ^~7EpQС3DCS[Yʧ?DDS aw߿)VxX帟AB}nyи0stĈCo.:wAZ{sy:7qsWctx{}n-+ZYsI{/.Ra9XcђQ0FK@aEDO2es ׇN#ap6dQX0>HTG5QOuxMe 1׶/5άRIo>a~W;D=;y|AAY'"葋_d$Ə{(he NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'P'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJBR_v'5n]FhNU˿oۂ6C9C7sn,kje*;iΓA7,Q)-,=1A sK|ۜLɽy]ʸEO<-YEqKzϢ \{>dDLF amKGm+`VLJsC>?5rk{-3Ss`y_C}Q v,{*)ߎ% qƦat:D=uNvdߋ{Ny[$ {ɴ6hOI']dC5`t9:GO: FmlN*:g^;T^B0$B%C6Θ%|5u=kkN2{'FEc* A>{avdt)8|mg定TN7,TEXt+`F P |ɧ<Ғ8_iqE b}$B#fethBE;1"l r  B+R6Qp%;R8P󦟶Ub-L::;Ⱦ7,VW.JE:PgXoΰUv:ΰdɆΰ (ΰ0eTUgXun[g, ׽-t!X򴱞_aM:E.Qg1DllЊE҉L ehJx{̗Uɾ?si&2"C]u$.`mjmƒVe9f6NŐsLu6fe wkىKR%f"6=rw^)'Hz }x>1yFX09'A%bDb0!i(`Z;TyֻΗ|ִ0-6dAC5t[OM91c:VJR9&ksvJ;0ɝ$krogB= FYtЩOte=?>T&O{Ll)HClba1PIFĀ":tu^}.&R*!^pHPQuSVO$.KMb.:DK>WtWǭKv4@Va3"a`R@gbu%_J5Ґ 3DrH_HI\:U}UE$J @ٚeZE0(8ŋ ϓ{Bb$BrW XWz<%fpG"m%6PGEH^*JL֗J)oEv[Ң߃x[䚒}0BOnYr猸p$nu?ݣ RF]NHw2k혿q}lrCy u)xF$Z83Ec罋}[εUX%}< ݻln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?tean`3-SHq$2[ĜSjXRx?}-m6Mw'yR3q㕐)HW'X1BEb $xd(21i)//_і/Cޮm0VKz>I; >d[5Z=4>5!!T@[4 1.x XF`,?Hh]b-#3J( &uz u8.00-(9ŽZcX Jٯ^蒋*k.\MA/Xp9VqNo}#ƓOފgv[r*hy| IϭR-$$m!-W'wTi:4F5^z3/[{1LK[2nM|[<\t=3^qOp4y}|B}yu}뚬"P.ԘBn방u<#< A Q(j%e1!gkqiP(-ʢ-b7$66|*f\#ߍp{8sx[o%}wS`ýͽ>^U_S1VF20:d T2$47mSl*#lzFP_3yb.63>NKnJۦ^4*rB쑓:5Ǧ٨C.1`mU]+y_:,eXX맻c5ޖSwe݊O4L)69 War)|VϟT;Cq%KK-*i ѩQٰ`DݎGu( 꿢\cXn }7Ҫa nG{Y bcWa?\34 P U!7 _* kTuwmUr%ԀjƮĀdU#^ۈӕ3ΊeBO`^}ܖj49lnAvoI "%\;OF& wctغBܮl##mϸ.6p5k0C5PdKB g:=G<$w 24 6e/!~߽f)Q UbshY5mseڠ5_m4(sgz1v&YN2姟d4"?oWNW݃yh~%DTt^W7q.@ L⃳662G,:* $: e~7[/P%F on~$dƹɥO"dޢt|BpYqc@P`ڄj҆anCѢMU sf`Yɇك]@Rɯ?ٽf? ntպ$ˣ>TDNIGW .Z#YmDvS|]F)5vSsiExţ=8#r&ᘡĩDȈ\d cRKw*#zJ9tT :<XK*ɤwoJarExfKB4t@y[6OO6qDfEz]1,ʹB֒H ֱw;SpM8hGG&ƫEJި_1N`Ac2 GP)"nD&D #-aGoz%<ѡh (jF9L`fMN]eʮ"3_q7:.rRGT;}:֪a$)gPSj0j3hLư/7:D-F۶c}87uixoxG+5EekV{:_d* |a%ĉUHSR0=>u)oQCC;^u'}8H0]+ES,n?UU{ x~ʓOy_>?/>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQXV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKdĹgJ8@o2k'Hr~4Z(I8!H G8HNW%1Tќ^?G(" 뭗R==9!nKErHc1FYbQ F;v?ob-ڈFalG*rEX}HAP'Hҷ$qM9(AHx!AF 26qxCdP!NZgҽ9l*(H Žڒ;̼|%D Ɖ`Pj . ֈ,ixp`ttOKBDޙ''aLA2s0(G2E<I:xsB.ȼ*d42I:<ŋu#~us{dW<2~sQ37.&lOľu74c?MՏړ@ -N*CB=i3,qjGkUտu6k Cb8hs&sM@-=X(i7=@He%ISd$&iA|i MiʏݸT{r[j顒x.Ƞ"m@Hy_I )j|s#RGI!dTKL&4K>#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhhdEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc0]rY9'z .(jHI :{HG}HDN`h7@{jnE#[dz;6nOj5~IJ|~!yKڮ2 h 3x}~ے4WYr9Ts] AA$ұ}21;qbUwRK #}u'tLi'^Y&,mCM)eu㠥Ѻ\a}1:V1zMzT}R,IA e<%!vĉq|?mtB|A ?dXuWLGml?*uTC̶V`FVY>ECmDnG+UaKtȃbeb筃kݴO~f^⊈ 8MK?:mM;ߵoz+O~e3݌ƺ(ܸf)*gCQE*pp^~x܃`U'A~E90t~8-2S󹞙nk56s&"mgVKA: X>7QQ-CDC'| #]Y1E-$nP4N0#C'dvܸȯ.vIH"ŐR ;@~y>Kv{) 9AG ćͩ$.!б~N8i"1KФ\L7/,U@.ڮO?mُa ې!rGHw@56DǑq LA!&mYJ*ixz2*{_;IYJXFfQ* 0kA".mݡ"3`Rd1_u6d逖`7xGMf}k/⨼0Κ_pLq7k!dT x삖A7 u/~&ӄMu.<|yi I?@)XJ7{ޱ?Q]{#\4ZfR-dVaz./f+yGNMGOK?2_~3\z=y}^G$*A! IcuR.o=MZ9zu b#s9@*иrI@*qQN||Ix;I}&ݢ6ɢ}{]x}_o>Mm8S]~(EX{S_uQ\ O/4D$aG87_YLTu;ve!҃=pnUגZ6p| G;;74^l{Pclwů Հ}xcSu)6fbM/R(*ȴd.^Qw %"=nluOeH=t) Hİd/D!-Ɩ:;v8`vU~Ʉ!hX #'$2j1ܒZ˜bK@*`*#QA 9OykGk,8}B6{/) ݆Y~ cPwίwX"{>9V0ټ_`#U8VdTtD_GU9V ұ{q:ObUi7s )B ۊZlzIA4S#x,T3ѱ ԶJ=rs>Nb: Q6ˌ?J%.Dl2ȱ%ܱ&6XmμQ ?J4WrSHTdp"ӹ'cJq2zPlX.0H!ND@UapVoGڧD5>H]f@!=߸2V%Z 0"G4ȇʩ@]>Y$ًF_Mm_Tt)ib+q&EXFu򾬳ǝRS>r,C2NfOjpcm{Ll9vQOT>9U}W3Q#vS}ll>ŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ? KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9UR=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8?|Ldl6o AMҪ1EzyNAtRuxyi\]q_?!zk.)ǟEu{_rjuWݚ;*6mMu!RgQWR=7걼/myanUn.Uqsy.>׌[~ί~_0n-rW3_6n=j{wLK-_]ȶM_ߐοFnnzΙk0hAϏa$ X )@VW)2&?ul_$,1=qOA;?U} 1^:XK \ )@+q(}* y 0< 5IDK5AhDK֠]4xMILv;Ull- }C| ]x ĉjlli˚| }xT-x.!>Zh?EV"sd!@БU ^p%pO3|B5=2怕nwRqR9~ i±za+HFNi>. EWz:V^&YEs5Ȭ N *7{!fRБBSۘ† Er/IGU}APQT]|XI X]FbKjKdO U6[3TTX)|*H'2U0:VunBl  `=/@ա06VNO8VGON@KgjyK?Wq1egI+ I.*F~L!Gf"LD&U 6tGd#fR*c ^tSLjnKS9 Ȼ \ >lr&}+̼d"I va,Jm_u)d靕َ| Vw85F3Liƙb<;dM-})C?Fw*IJ_3UG'+¨[9| >80\+ xJpΕ`p~mg˗%F Rg(6=/r+%a>w Ohght uЍaRs ^d6GXAf?V_mW puȇ S:tŴvŀU#-*mZ5k5r)_x*8ͼx@(k:_TX%[paRu~}#Ѥr %A%`;MxB[CzR怕#H% }8@*AM.SEhd,rKrʇ)br\+! s1CtӒNc_:F*`Nv;ogQFa2V%ZniE|nZ&-I,t*ώlo Lhnٓ'Xm R ˍ-~ά}hs\5TT%;}]|Bޙǖߔ 3\ a-`slԵ怕e7ːزoW|A\Qu&'9~ l|`pΕ [Q =r#vQu0 M.1%]vRat'IIc(Irw~Z"+A<sX4*X FVGA<^^7 vq&EwQű:؁6y\QbR9GuB/S5^fa;N(hz)}_vq@nu@$_DVH|08W12e_ʿd{xlzUܝlNDU j>zƖݗ&!jC`@ qэ-V Rt2m%K6dX)"]lj齔{oY:8VmS!:Wh#O0} :OVGL.xllT_oqqqLec2p;Ndck[ Rh6T#0H Q}ppS@ώ@#gƖ8sѹ e^ CZLu+."T#yrHhlكʼE-X'I^=bKߙԘ1"+< gb`[c1髰?(o$[eR6uOœ-m~)-&>883\6y 8V -qrG]~.3jsqY~ sjZ+9[rAJsT= 02ݬf¸9Xe>sY~ ae9} x* zjC.5Wg󵸊y!1U:pU!ƔCm-7^w]斻~[hW$k sE0ڊSq:+EKٕ|dvvjjy6 æ/ML-yz,ZlQ^oAn-})xǺǍ--qcl:WLg ӁvJ[ǧc~Of+8qpçco#rCtKӫce0!Y-+cxMK-H_2:Uu*corD~@N`#mR:?歼!IZ5>H;2ޤ:\Tq]_\_>e˲\oUQ\W?47Ww?KpwSSۘF,nK.\U?VEuY]^VW0R=<ު˜˻ x}[׎'|;c^ M77 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U}m*2@"G{yZ${˪A6yq>Elq*E< NX9@: Ih~|Y4sopp|v1f2춓t$]ko8+F/+Epto6m.p)['ʼnן$;NGmݨYl_p8#HIRuGI X&U"';Jd f m$MbJ"xWϙ,Z Ǫ OH` Q T<#ʢL*qWؔciUQ*rp(RF5+æys+8+nϒOyRH?1Uaj79U$x{IU)qGHh5˂zFF*9 W0ձNMM cUX&?j1>b,|KN8pLLcXӢ^")$vg}Z$^BfH2~]R퀻vI n8Q; ^_Xe/j[y|,?v|,m!heqxpczY\Efl[A|`mqXE'!jƉC1Wfܳ }k0nXR+F!zFM9>}=>G],yߗ/bzz>hT#&`uONx2mՋL/H*T_2FN`/HA,lG ǟs/ S};M], $1>yG%c|Z&0[|(400 Ji0}/Odbt x.7BPUl:Mۻ2+۽2ɕ\91-ꂽ&O<)ӨG-(u0E` OT &Cv|;O!OW,KG"kGnǴ}"L$?U衼F1ScUbK"aDl,?e+K []x)2ݽwT!=dnq0ߕۮxpP#6$P@fQz]zqԃi%Áw (,%>LOCm.J双}fŮSJOVd!uЍXp״\P1 8ꌚ fJ>0[OU2'~tGJJ&qS7~ ^^_Y9YuS O%$cspn%`dNYߑͣ ><)V,硼tw3 ٘t7Ϧd?Iw`:mkr%%hL- V;A$O 2hDFy-bVr-g.:WC`],ƫ9q̿&MVyޮ;&' o'g1e͢k $tک@:S͉i*dy{:_/ߖ ͧvu9uD m?hMZ''etCsToW9˒贝ѽpr'% \o[9}|>ywrƤU KArqvQez_8XkTBK5eRN`-E]:yƻSx_œ[O/~NRgXd8eCP%ŀ3mR0Rpk/K 7eE2aJO$d!/CpWAlyJqYq g?Fj ijFa+p0+6>Nm!ehO 4V|QWun/I|rlG)O2| UBuKNs :BNoF=@ɼD4Y?{¿KBDw.l`/ Tmk姤!HA0؋&Γ7( X\H K_.2D%2BEKm3u)Xvk`Cā]$y.8>`T\)_: #j&<}SWSǁk+[u]3 oK$r,l Wۉ0~T'ApPt:탹k sMDe_)pPkvn*ЙywzUqks~iPL@nk+C'&0׿]*|VBil-L* |O2HL:@ntk!Hk1+2fX錅H"}hj' ((Ằގ "Lj_p6ljA;ݶ-Rm=r8"|$_{1"Țp J2 wW!qwXr~<{{fD e5$)EM~}0j;>2GjcjhBm$|ɦ2Q#{(n"}#,+FdI}|dHYpj =Z2nD(fiu` Z6#;3+`QȲ$av c>}ׯhEaEXh! g{cpB{fE)XٵԝJd6SJ8ʶ17v{M泻אh/ioI >;muhot=W# :rG FG oY 8:c Bu_`8jQ 7**}hj:cx5w%&U8$k/XS:bZh)g_tXt2dywwUr{ȡR]2vSsV*g]ߕOϢRi{p# Z'o Ā`D:aLM=e1gVI$᪻T EcDqeBCw{GZۘ(ù g8v=z.UC ii=#ص-#ŭ3"W%7X]Cap~ jaB i.<݉gϞgD;~:vcYF!naU o)V70lѫ1dZjIX7uĔC0F\+F ̣.0c!J\ Q(.Q:d)8Xs%% eI5 Fط5hcIjxU< G\W[WeI6C #6SKGqj.V ?idcD𛼱RԲ0Hz.瑱H ^]K4R]vmC87Z$Ew)F$UoѕLdbzp+Buٴ+"RooQ 0][{f˜Oe k`9R2=b7inDgn/Q늨X]jxn2+X+#dOh;xu@iHՀE0!UK)qTm_D$X8E+~!.\yT-umAYr-ny&<0Ĵ˺w+t8|VmfVwϾN6 8XVƮH&\*U|:1uI,hgŀ;-+ӣ:OR8CFLBUo(5DUaYXRzٶh#yWʆ#۶6A(` 1,WՌ? kMCbd t1}6Sm"t ZvwWj_<(e*FNzmSǺק3Y]%\D)Cov\fk2ڑDb"tZlVut)3TECjQ±뜥+$7p.xuɎ , /p.a~5+!m:j},*$z7 TCDCtn±j߲(DlOmR[ Qp$uM-pC VB&lc*^EyD vp@7d$=#.e"=u?-d&hO>mu=. Qe>=O]AyRV:\ Fmv'I_6#1-ii{7Y׌eZAX*] wvz QJD̊c1uIinR}0\4i:D%A֕0:g|c7WW{G=~@X8>jAԤ@1?Pj2HHWrA,xϢVq]Ӫ;卯,ϴv {_IZwzh<)0"CV*Z.QBrDtf +#\U$ՖMZG^n ~&nFkKLDv̨fYg͍uo#3MֺN(>o&{AM|bܕ]ħE"__#PCP"Yv*>PD:r0y +wp  ڐrޚPDA@.%:J`ٱ; t &i}bl; WhH<;KIh(;$@ p/s& } 9>$4_! d J=ʾp;F'FWC6M ۳[)8 1ltjᛮy h"(4|ޮr%=}^(;M>qo: Q[f0-}˴ͥiZ}P: DӷjmzD/͖zܳ6bW80P"0`3,B>kdr}ih#uBu~@Ӛ۷w(O} 5dm~w5mJ_puǜ:NiI{3MCF"ud~)H^F?.Їd,֋~9cOjn?cRB[yn>gyBڛ\ڲ]ab6j5U;oo;\64<]n Ve@䏕f{چճjU[@9t%'4/.ϏԯwG'?K`Zg{l39DlNܴ}z`!K&nlg# #DZǦ)b:4G;sHEnW&{6-@} vtQbCcm@qG" kN׽sc/b{pGb"{4vH  DֻF㡽Z@a}Qu8kn;z t\tձz߹`J`qiw*{,L8G^Anu%QW5u]֝ u- { qgl168=rEnGbF^o t22 n +sw[gW/e5xJtie2ﯸdjruTK#N[6+: s^| k;|^lOayT#[xޣ^&).̯[dY}tDU ;7W 3Ƃ< u˩M0-o +t2v*,F0xELeNZOߣ}E0 owv+鈛esnWm 6@;Z߿KWÀawn9f^&kz; 7ܱ: P9}{ |]HChWxv p:-,Eg(b98]@<~Gb}k='5ۡVn3<oABPGw)QNV ^t H8aiHe:uAXdEPI5G))rۓ$yvP0MPE܀ƙ[=yEpP#.@b-s( IQ|H I>v"ECܿ (|3z;R| …I u#y & sd C=F1,Ad&e6$.JS<,ɛ8ՃŴ(Z%8 J$L 2,11f)5M^|&i*wj轅"Tw ^nX++@L^KV$UpkT }J)^x/%J5f>\f(DHmI1d=9> {3"QHZBs  ~2 Z2b!~+)kt0,av8t8?;o%4A=o!)ӛ 4V8D&$HfY>j`qUɠfH> }f ,zSUu le)ȽO玵.Fn8"hB,(yVy\ ,/إ P$#8k$ @U: 1R O ~k+Lq[mŨl˕3LLH,(Iq[d.SQ(X)n56?azx+Q}8 ͈дÉ "p('w8+295g:̴Iz,(qr]wNp'5@ -x@8[uH[3vtCMEB~;ǨK'/E@8s(Tɰ$Xeq@9P窊5^].R캔KX@Å눁0T}zRQ[O r O j-&x?;O1Y!l> ƄH_F1,}Q+2>; s #K8/N_G3?`^1Zr@/%^+OdZGemg8/У,g`Y±M .7'z|~41&WO\ &@L#~ʦaWH=xpq9Ey%¥^dC ^ߧp"{#Z@#YEbfg ɯ$:$ /%DMAa&1>!`0w }"CpȚP8Wqn /ewŻ0XjjUwP#Ax8LɞQ@ٓm_f+ t3|mn Zp0[;nt)LXI῱(Ҭ&ۓP\/i>w$Him؍Zv}inݸ.oS7 oN+|KJZmB 6'z֖Zj 7 ޜPe[jo@&ـPgsBlIM؀P9e-  u7 ݜPe[n@&ۀPosBmI~PB _F% BteK ?e}F%&G`WH5zcH;i ˤ(IX$Hsv+!qӿ | /h)տ{Դshm>{YRм?SrLLȯi6,2Wf^ϷaW(ytq:r*.]eӪ\42z6MG*В{? IG}|$<4p?͢~b'eEc4Q)DحZ4o` q$R?,!}]t h~fh5E|GI4% !=7n4u80*GcIRY^9JAJo5/3Z+?}`" |U82nY}ُ5ϟRF@T[ᣯRr SU2 g#h*,2XղEṻ[/I6̋(vp{$3_A6V/D)C}6'E82nY=KXn/@g62|h@B`/O6{ N@lݵ&>/AG=)es~7'g`AvZ%\^W Ѕ?驌Uֻ'*H>XtVůEy|8ͦM^Lp ]¸DSV_j<4 YyjMʛ^J\|rFE x2Qhi\3Zjb%Ldi L GW*nMʂC.I8=\vT(MrkCj=+]3*eIbthܱA2 AwĦ@r*D,E^(ifW=KOVT/K?ϴ 5GY 0`(hLHʇ,:ꅰS %W*/edO*OCt7k` Vp7oZd L{hJpJ@ %Hq$fM  y`IM8 H# '+H*9&8?[2AyҀ *H"wJw@$S_)JGX D9õ߬W#J(j<3)#-ȳb$s]794'뿇agbQkSc{Ư~d,j|3,Yv {74c׋ܰ2uze5rVc[vjN/R v`~<>k`(Xz \__xoA,*<]UN-H0G!p'r'tJQ$cu;㙪Th#NFgViǵ>/> =`ym`c?+A,IY6/Լ 5θӆ K-0$*ׂ;ēQv>P5RhFzXbQpa9X.!4Ԍgf;N1г:<ݴME.uVF L[ *C` z8kbPԤ))Vm;Dުw>,RCii(U~ U4②HuØ~.Ѵ1Go Sibܲb5U+b3T>7_0U?p_=TV^ ՘ykP.7*0Ԧع,,yl_Y t߲VA]l`vCԲ9+j]}Vsgrm^Hy.vu\VVWdKzHGnh*5|<t)̂r,V.dm2UJ=,sm*on  y?IB?vmwsݵcߘ{L[jzeWUقPQ1+*Eb*k&Fy^L)''y0;) Sͼ")ޅCԥ5[M!^J|RHRGtjtgs.jUVU氺-~X^g"XUn2<ߞK"l[b Nʱ@ (HaS>6Вp8EQLӿWBE'+WƾHS0e26GlkL\ KXk@g 0/wd)dx'yp)|jpq&\3Fԕ4~&3u&gٗ^xz8MLm)B]zK,875>.Dgeө};z1-)翾|%pd{%¡"AץOA7!dt #F;y? JJ\ pILr pg8]Hpʍ!э-\̒pcy#9ΕyJ8j+'BAu D`PȄ,F-n(pYS ErpcPrFpa$G&FUwrQmIOtvyeMd3PԦ .znIهE.bR![:(!ÈAz&bF =v] {$ ܥ̹<s4&OpKᗆVDLo(f41\7܌Fmu^!JIt1ak~3- FxIc-q3&M! [.7ކ/qp?K_a$]|!1RB)7k$o+MmLsI8J8c(p=J )^r.1F1Tj7z=vgpZ"цUgMmCze~VhO{]y$6hSw.%'{R [ eÍDy 7 .eU;mx#A0 L~zLE+s&Hbnӵ9,nNCcaF/f_KDJ&VxlP1^6fU3&'P&Q펝➬Xi?ZC ʎgxy3hG-1 U%TǴ%d[*DL!fH]YV53 v`:3CA(|AQ♅^WʄYB;F bU{c0Q`9[:-E )|2@j}6 nC]̯0Jk]x<0 =|Tlc@Ha(KXtCZu9'a-D 'cLt%nuHIK$FE_"1mT9dÝYY ٳVDf HIqNKn=|YUOVE(ⷋDr?KVUQE{4F7k?aj;"AN k3ZUIVopO5HbI77[FE` ^O]52_cb;,KHpݓ*xf0= ] [/NfQ&:yŏ_n< xiی Fx4C>f .q I/ VUx!AN6߱ؒiEp2ю -G &,1sNͪԧ 󴹋S i/&fQ fbeNH&9ZZ9Eǘ uYULJ|Y=z~iU y%*,+e1;I54 Y2&kG{eO$8zn3קeɓ.o& uQ553XR-3QB +2퀑tIĐW$8;g\U:^2 ׵DJ &e]ؽAvo`j2QC#wiCi3fiጓx^3Nw0"W4 w78pK=gޘYϋ=%7$8:*J9 ب' #1J՞DGw6W6#w}|.?l ٺ̬q<Q5\Rp߀%G^=HQuMZvC&㧏qqIقky|;|(vF22U +tPEh&xAǫ.mE9-%AX;b{Jb/Vpba}I* F .QKS:%4x,N-Zc)7s{jg혯7w=~d;$w{XɅ{6tӤr죿MdE.o*čh`{u_gc. ~V ޒǞTdVT)ƪ ~`Hu8z>chi+y XnwC~Bg>wB$"8ĺK}wLzۃ%ho 1+gbm RW4n8(`#/E[T 5bЃ 4W:M 4#M+ j4ԋk#6Xx9x`WDzKQ489˽'IڃIG2R;%Ż%9%nek-)AWRCYM-C\TO\([6lBa즥ZW~h8W$  W!sVgLٔK:o[↑ClcVj<~+x* _nLmanZhmò($( #r>8[t d΋zU70ewF*è[_' " /|܏^:}:KjaO/\RlAkgʜL` =3>Y'@́tFie]*2ՌP e>xIp4rU+0i˜޻!L#N LBf]'X=R6@FHP b0yTmA &/LD͉\W3l{v$8nZkEZ+(Wգ5côi] GFԃ]P y1q8o|AuԳnsO}ѳIpVO nw{2M6ColI$ w[Jp>i+}lX0Ajg,t vp-t$I\6SU5̂L-]S(ZTɼ1XBO+5 )N0C7G`J3Q:KOBYajnD㫚nj^d yz`I 88ƨ3@(0ͦUM ^.XFerz*7=7SL$D0^3 fc΀)TMJ.V1c`o/c{$81<],`]xGŕIxo1oiheqIp4L[Y/qf~i7v)-0]P5^U5gytͺg0 }g$ >}~( k-K O o@5Ֆ+ʊX}\^<-~ #5 |d 3zC/@+dC.p̿m{k/<3G z^@ R *s4Td1Fejm'mxz9$p$fd .-q@Cqerq}4E9# 1fCY),w uDa$Auyܒ8xq)+(B͋>r#1U5[Ϩ1`+aZ ('F] |"[ezx˱qсߦYm S=7kjB .t-eY5@q2C_հĬ3/AwijM{r$8M'oql>X08[Zfe0e8c9>(t5neL.Gg.^}PI~\뇼My~x/\97mi8%=g\)FZѭY 6ms9\ײG?{۶Bv]?ַ(o76H)l=$wZ xw>;Q0Om?%Gz?kL^;W)#X. 4YO8/E'=^;2Y1M1C;"yVq7fo';6R^5Β(䳳/}Af?Fo׿P9cA|F#$Ov<88{Hę&H30HbajbԔK L}q>^$ħ=mDT-S@?GU3'k|ܳ@3sd{.o:ǣ쿥-oт<`s𮲡^xH~\w~3% 9Y{3^o"^aOY5nU5֡Z_UUuq6;5yqLC `]nMwUu( {b:{|/8Zwv| fW`dz_@ yL$#rxeZ~X'd=7yta.9q^k fj/nV6oZ߃dW֝ *# CTv]YGtC>+~7rZ18H7X(ksk;rӥ+׬MM_#0s˽q3SSQRdŵ߫4 @j Ko\]{DcYCe|J[7srvUIb 3<v,9?GxuطAƲ+Kt8 njI>.]r!;8 nTUcs^ܥZ]{L/NzL~ǟyMs\sUgUyjek_%qK:%u3!|fʜaN,dW2g -CJe#7+ITtg[ bňdң䥯ʓ2s?cuJ p;e;HSosH JR5=6w|bw-jJN)KjOtIZɺS!muFwI2N8b-?APUwX>(-r(KJD2Jofnu; #;59b] 1|w {qIKyaY&[\@$%. 95/XAkM23+A> R:bZ>;.TFMc}QEòocx6=hWW3Chآ0;gv@`١PУAF"6֒}7'^_W%lwg#sc(ZUYedIgⶤTԒSp WpYցEHOoggó+dUK72(|(Ϭ\هOWt^8H^yFriR[oJjjϪb3,#|6GNiChc@4Y@u|T-u"F'`7̂/=)hŪ??[(9U;\`DiZJapʬVJë/2l࿪T8!|\flW) Нvn? X3R:s r ˠM. 8jqaz sм%!v"bfpf S>DFHl{,Rq⹠KsueW4yYA'ʽV g28J Z0<BРϪ%~AZʭRLnb^/nrkNY+ѾkN^u۽gǚ#1.{X7h4"Oz#\t[? z AFV ^ϓjS%FFB5`kTCA0P<wFz2՛Ott`b[Ⱨ"zk`{u=ĂOk"j?\C\ d2_O7<baBd?|381 !PW p^;=*;[c8i36q2[RX&Mn[܌^>q­Yx(̠F'n;w -q(aG7xcAf~.ش1&l˘@KE- .ƨ+Di)݋[@.-SBC:ʂKNRZa\SuwMMԚFC Oi,GIw<*W(|`UmR'IǁQR6 - SEriq@#RyQ̱ѧ~uq[Z̄l-{Z*=bP>A"TS:-R'ye^!#"3Pnm;jm;o 5oXX7m)E:LPHH*}2=h =@:v&Ƽ r/R;iţM,9k<-tZ}M}>_͇*BML'pyt|0%:"U\#Zc(+%!X@?/'7QOϒԯC3#`wU|Y`Xq~kd;OP_|^]^ۯgL/SS X7.蠶%(prG~}`K >A7S3a6uaĤH#pFg?}|P5/ *GZq^$TUGy5|쯬#•.U|L |6}j&nuŋ@* h& bSÌƅ`p0|k>BZtE6͉TN4"TRҽA:PBL6 UU <bsv|!Xc&C+s w><`ֶ7Ůlv G1ɧFB@̫Œ\Dݠ*jlv`XdSӝUlk* op LQїl@W؎;;WEpomÿg1^XYZ}689G1վi8f; /"f߳v/ #KT7R׍#Qܝ*\Qk2EQ}kYkGv0g!4x5Svo^5oc C?&ujbrB?\OTz{.^KZF ޞ僠dPr`D|J!|zJd =OŅ3A-'z+cY+`?yUs}X^e))_%%6}k-y?̪MLRdocMC$U d20#4Qp>^Fl$6WO֘l: @xjn؅ſA@I2[*}Uw^LB_|_! ywOꆴmoe}"?w]hgJ9B'Oi`G jR̘f 54 1_t}:| &T"ߵ6|)\;[OQLs׫`ĥQ4ADybZ f)32dUT{iSϑX^3ʊ"<V`9Ƽ`bu(zƂkm !Zdt)qs KXα<_ǵHm ocE)zhN`Xg=9ekPJX^\hq'.~x9~wqXk;\=A+J ̋;gd&Lq%ϓ^ HNzPVd=U| % (fL.SGa;ޠ8 D#O2h*|O[k$Sp0F/%ֺ}Ir'FMXGL|-^&'&'&&#-t@i!k_ !HÃ[-'Dpp)#Bi 6iMuPp=nT|l{ QGC<]FR Aqa% )WLw)>þui%Hk !Th 4ds401˨):T80tH$ٞG`8';8X4)BTb3Z!Tj@Dٻ6ndg$?ׇחz IQE~-dpW?Vl--+nIjYZqC~3YT2x*G叴V*"h`LɚT&q:SV#c)q$bgmfA$8ӄ`\Ǔ^D>C\2H8s6MeneYs Rh`5*q۩[},IdY.!)9\ qJc8&IJQjhF p &y*t9w.JʜWĥD0:w`T (a\Q\fm'>1*I3Ƣ^\LZ> ?C͌q@ #AA3`T${&Eq?n5A F~<ٍ !M#}F۾:^]3NĻcpVI?柔)@j3;,#?/@D' oVYiV24ϨJ?Ag7柭5e=j*q $6?L9mo4ĝu*G'R!bjpxdI2չ́TS ybx莲7 C{֗EG|wm2 $ 1Y b $4 O܎a7矶iM)Ti Só$0 ? W4 V O V4a=M9%S ,y40x`0r'0L荂P3=kR&]`@8#<";V/ixjJ00`TC6HL.ዩjd4*闊1@qv"-c8(NKoOunN;Ku|p9K+ ^/^N>zԹl-;*s8oü Ol3FCIPM難8dՓ<ɋ9#4݇&Twnr{yS<}T = !B޸߻1jCc'SNF.=ܛbOݏkBv1~xvs3͊%oh= ?թn{?z~v0=j`"A?@4F콸A}YtlУ(~ؓbP==WދPF `D3䗕_]_N#sdnBY|ykV.IwfQ_?FkMWUM[q+P(]?mֶ^1" *Ni ϲ~>xq |;:>N߽ln >9>*Ͼf Eޓ EA''eW{=t5ňjU}!s=ώ- 'ŭ5o8Rߗ6c*I?N>OoH^nlS,KӥV;jBˮR˽ԽQg g!B||~*3_SQ'΍P/b^d2[4Pōy=}J|8M2IiǼQ~_?㛤ӇW蟿L?emw.`:uyg/(1zoL'2 -:ԻApV=%VRן_ Fk}0/8; mƓHhc ~B ӬoLN4%Ry5D0aWoyI&oǝ^8wn4,je+cٛkGNJMan> aF 2[~zҏ}w̤2ffi̼n 7T.|t;Sy4\m(Ԝٸ?a<}40܁3OC"8^rToNj[[R^WN#MP`x؛f܎{#7}4\lHy?ƿ?^l2/ώ{T,m`8С&[&"!_YuKzK Ԩ?j!V`J-$H` -$n45CyDLEƤ Ajʖ.x)YC8DL{ 4W4U6YMo\rRw8 o\3]Ɲ? ̟(O"Jթ  o>C"0'0lY0ipP!{E5rAEō?݌'V21̼0^7mao7%MsU"u{LY\gtyvt^5AMjb J4Ngr+{\gjSq|Zf1(H"A@/-YI8vl.x ^FlzUcyK[Pyw SHhETZƫ >(.m=PZW= WOb+CٙU,ϔ~P%t %t9HcceA$=SE>. iAVdx#x1&5c8 xy^(1_I@z@;__cJSc&g:+lʒd)5`YcIЈ8 0 ۿ/IM+h,+rrZI٢^DJ'~FvƨP!G UcDBUs8TY jde~pq@[lF0Ik:*?=0j_K(`-BYzH&j-Dvaoa/#-+W U) P*T { {=LFDT}' {wzZj)RId]ۅTSBlRD.PEvaoa/5L{Ujƹ큩"Σv5ZS:JalhTΣv5JpPnL%DR|$_eQ@љ'hlo(A4jH眶Ǔ"8tknmڔ2! fP- ۭM\RF30y TvaoFRN Y R@F ҟTpnV7ߍ/ H=i64E&lrz[LrS+k~XfL.O42ՙc.&ơqr ʼ( Vi%G6㜁@j/S^5c6'9_E0i3%L'dNV5o5ҟFR}!gbJ/PID PpZ [=38c dB;FfHPlIU%~6U^Sj }D# I!d`7,%ȒYTDt˒v rZUQ5 -*Xj#a p&%6!eb =Y1_>aULVj*#q}cfbI g- L4ݖH\XiM$EIRG'r(_6DDV#)B4 :RTU{]Š&aHkEc _2r"9_Hu4p(m_ڍ/*22- _cjѨ2 6vFzC;B4v/.͕I-aݩiyGc&* HRAF[#o^Maծ95o_ŝ7PU›v#ፖ}M}L58bS_xӂx15>DnVX ,@&Ho c)l y -8Hl|m:DJ( 5(Ԗ(\FC.ȐEvg5Ų Բ-hVFz(+YD;װ$xv0¨Ex\EN1yRdkYDtZ ,{=Ip`͖BDs}AFcII\Q*J[$P7I@jlN dL=앳(bGt϶J+qP=̛ t欋zA# s@q_Z(]f ͉L2&O$uH@!4X:I!-FC:a $D=p&D(2gN1Ky竕Z(66c'/(଄+2Hɴ o/e=p9)d=%)ʠ/W+ d7@Pj>;B:DHzlL AhW.XDP_E@#F@+ A6Jp%G8ɋ+fJ^CӰ #5qѱr}r2(5gwSd/~2ܭ3\VZ#+Hot Q+)q2g x m3) FBIQ=sNJT)$ʡcɡMu mgJKR;8n?p)U0$ܢ "UZ|7!I(ǧoFz䨬R6[*4 ܘANʘZ|*h(IAF䎗fQ. O Yrպ"C~ P (gcU dJ lT04ʢµ'Bl(Ι}:K9qF4%,i"QI 0ΑO eD@M bo ed19,1U !c-ϲ v# drR,] 7Q7FC΂  P<A+;e(aH7K܈>^[[F–{Z4cop$W&P ~O"G!pބACȜ^q6 Wj bVHK!,$eX{" ټ?FJhPRtބηO&7e m!eH#Dg $F>ƴ̒"cہQ&h$J/*Jm,t8:dSaK[tZH\ȫHRX"+Fz(~dX5?_&reǎ,JYO@3W(0xvұZ-o,A_P[k !#} h,bcS@cm , %j?P!2+ (V*MP n$4HQo1h0'HBϐgeaC9XwZsrڱPn7қP{n~.Ň/#M۶@kVTi@s_Ȧ|m4 ߍc -2JdAvmyOIZT[󆯗{+aNW.|~?'G\]ë]1okJ m1}Sxo}GGVӧǏ7||bL'?vtJCE@n^8ʡ2ζV=dz>|E?Oe)@ 3V%1c[.6|_0:'lG#4(l2 ~V03Q_eSPXMqKvC/i oÒca0˪Y({*مë/7CՏ|!O_fbôP_% 1eǂY̶wGSb#fY"EQ`rgx$&g+5j0XQv0Mz\NE0+F~Le6F#>ȧ`+ìe4:sx]JCw ҟ 9d §Ǖ 4)H*?֙׶f}M57WkyMH l?~mf،flVr6Ħd#$I֢YІůKxݎg׻`f)T,I(WVvJ@77lcZ/|3WY=?wkWķ$/Cݕ7zǯF3^!u:zog yoc}G/A5emcyش(o9&S0`Y&;>|_gVȾegKd6\؁h*ώ[7龕o9w|B7PKrJl.\|~fʺO&u_o>g|u뇃;<{jᨽ rƒ4ޟ2]~j^aS7 -e kIfk첞:hʺ́>|?%r Q$լAqe7 ʓ HF"Ls?Ѕ\uUdV}Ry5P8 09|7c{Yi6޿?!UDϙ*+vaxJI7V7C,DPWVЅ͓$7Vɭ$T)Wf-%1m>i2}ƎUU̘o{ig^+tL;I4.vn<+oY^'SŤ, f~J_7/E:nCRއf~*v)*OGўS1`PKmcl^DZŲЌyTy1;H?С> "Wqh1-u@l]%Eg75!j+A>7*W%glM*Q){6o-D_PwZ76ćEeO=ˉ%9^xvWSfe#US1'113m FqUx}UyA Y8f>$UjpVKwTT"jG =QNMPy'qg{uǴ1);xhI+;WUj|%/w1/b/T[R}5X/3a~k75bF\mf[σ˘*Sw1)/|͛CN=>JM|Y7åstutWv-+]iJj9_bp;_+/rƌTf qxV; )ԫwx뮒w]F ٻnFh4hni`"@bײL=|'۲-OQv1㑞sKZ}#H*R] ͟VgvX-=;8|s6z!Ĥ̻FJάH];kWw_>]?ϖCݣi c9N-:]nIC-C|YmB;'L}KbOikv?rL{/ 6T!Y"uXRZ%`+Y >n L}]Ut|OBpї:ƛ}ݥ/_^D??^s[`X; pڳy+pef. 1 >~5ÕZ$ͣiR촙 sP/7G1eߦ0+6fs~xsMwwiֲukXΠZaf-xQ)Lđ)9dӱ *ȎjT_=հ /W=HBB-> XSPVSObF* G'&MmpҬhSHV|t]E?uBX@uuP|_bqH^`B~t b\iW+c'Ƭ؀s7ZqumMl=k>w5!pa QPz ~\z07Wg?עh^~|`u*;}˚ٴ}J=sn->a0M땃lj#vtu{HU{.F|mjQJZiF;+7sTLf*"h!0LŻr^D=O'ڹn3{Xkpͽ֧U/ڪ}Zn;"Ԏ+xoI2 04 oFkpড6C<_ӹZRrGB_U V:TdvqAVc'6]!H^o; ɝx#1u?yxoo?U/@ŠDj'Hh?]\ Lkiwmw\;kBV=|6BhHP«zBT]eY؏'rVM'1нGͳAND FVP+,L u?ʆep <Æ6iM 5XIMonnV_uBqD L=h 9hƽy Bl:8WԊzYVu-]WUq!z_T K@CR :4}rto_`v\kk7qfc-,Wma_0 }:YXnȱ tsR5 @;CQa:ԸڹPvL ׎D[+\h"~[z/IzHӹWym5)12m5bkitfN}=M}$ᶞPCOZhP@+ٮ(vzf@IΔܳT1wjqͥJ9bRV-&]X:5k(xw=b[2{=e-V+ڗ+ToQ:0~6|!6OM_ vlYͦ|f`Iq}%Ta ;b5AI17b5AIxbd\~e^h~ؼO:=I Zx6RRD"Dکln>9}j & \$[b􏒿liqǭ\ QdZG|p>t5zziBiK*[RVqj.>P4Eֆ4xsI%F5VEfI!3ԭl'|& xzpݸ+4pRWZ,&4&kSxljhi'^|hù㶊gU4bjv⹷UlZ7l6M0lt Xm  # c-6C G}5K>NHw[L_|Y& iVZh]ާ<*BZ)HQD&*qEKh<;UB^ REfRaKT3l9kbT3lmYp;t; xRj,#.kWH oMnrz>4nHДi.WmGVRFU {4O鑳3%N;FE]Xtw&\_)MM&R2Ml$V'We:0MW f͔/k]Ф\7݊K _5>073N~P]}.aA˳i1sGL(y"2w?|oGď= }tEn 3J(uЯ1+8ZZ{o?F4RvzA'ݶ^Kfl׵ cՏi)H 0gl\Q݉kC Q0QԡVԻ]t/Vj zww??{Sg\tˎxŒ98ڮǺG715LsK%XC7mt՜awz? M;;Ɦ1[BS}Y FG7/~~uհw7_ }?uhܕ??m4Z8hj͟5mΔfߞMu/}G?˷ 0y'eϥ-6z!%[Z) Zm=H^He<79gx:\?{W"{+O tǔ+SdQlK0- +A!|I̗JQ*6E1"fdEe@&%F:IXqcU*I+ ftc\l#B䥤E'IFY=&]_[ &8t[}9Ӓ&S2phgB"kІ*F l 'xS]aCֆUJT5%b4m*2t?g62rzcxf +^C9嘓$幏y#ZZ&v^.xmE6OΛK!PU ˡH͓>XT=%6+YVGsFUVW$FG;|TccZ1{ؚs|!%2 Chqj`)$$e`PY̕3uH" 9ROIlt^"Fa%K[0 "LrFL0^ĴOK2H`%Fc gB^Lʘ]Ryہa$μ|.VN,kN)7 VA1Uo.CE\W3͠0N-4/h;-ra Q7Pːʃ*tėY "` ( <\c0t b vulPL+1\d$/ )D+em`DF!d!.`)Ii>1TB`V!ڰ6AE6ZeR|NPPқ*BysAΚM-x XE%Drx*I0ϊ$$waBOJr YKdHS}f䩈S P$$_Třd, if.A Av9cz@Js3.cH?g(c)VYP C$X!AYP*+tEK"gykcqk܀ AŘ0 cY!!.A RF&,MNpPgJZ rղmv++_ JNL6"L9b S2渥`ܨe%3 o/ 2Գ0RAOeZ\.S1..UoVhFUDI))Ǥ.3 %$$+@qDTp^7 C`\;@Hal x">f؜˅ Lr+ ^g\{H3W}Ji  !;TWU[ZS4V(˔HrMQu9zVjZŸdb)$1DEevN#M^XߧiQҷvטxYQ\_N11].X:ئsV 3 x60ua6fАJ'j-BL\0h1# e 8&!q'bE*)tqaBzSJII&|ZR"b!}RV1`*>:(=ɋ  ׊ Ct+cix8ވg0m8ڌ4E8 G'Xż,+aa0l3B^ a; #}XQy.~~}DÔu! ֊֕^b,pcG!6^o~.-‹9HQ1Aޅ\u`#DNwkR D(!}B,^Bn! U[c%vLF,qhͥPEmjnJ>h̨F]n0hZ'Н)@1"eQqdj 8+kn o#0tj$̤8,&#< Z!vݑdDh$byw0̹w,0-Ucp!Ht"$UԈlۥF3 A[͙KLE֨0 b(74RR.[oeؽ#(WMwĤ#DK6$/^B%cecHtqmi7i+wɟa'a۴$\S2%&P`ԵEwt ÚIɘŰѩ`*@S˿;Jg[ 7^Sf Tvdy. r3c\@ʻq@/60a  L 땃ke~ zL%aI=N1sJ@l,2bLmFuPgR"\LA >fd C5Fi\k 'k7~kPKp^%P6tB%ġhUZ#ӆ D oPX 3!YV2!С/D2/1% 7a 7Ȓ1p|;;ž2TQ+PP)9Po%r&x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(⡈"x(6e1fYM<[VUjx(y(=CE 9CE<PCE<PCE<PCE<PCE<PCE<PCE<PCE<PCE<PCE<PCE<PCE<r-XI9p[ :b `?ǜ|I@;w?YHX̒=8ϣ'n6/RJpvG"DbPjK%D,F{梟3RxV?~nU4 C 8^|ݻټ34@6ubK|%b0[,|#k<lUX[F[X{ņ`zK3&D(PDd\l w!DaJZZֲ_Xnj8^gVaǬR9e YDl9l%bJJJӕUU"VsdkkP@(V0@f$DJ< b)bo=pԵX fX%b7nq{[9^Zjv{{V0[#+Ӽ;am)֝2!l{\Hd%b رH-;8Ŵs2zyW kMᆟqϗB?ylMZYg] CO冽iPNJƁZŹFZ3b\CQX'j`^x7[e̺JrɴEl=D\]FZjYXJe,+[.pY/-IXNtjسR;k*+îr V*f w7Új%-0J1+r2aoɷO]-yw#*[n:hb֪JĖ3L-c(/|--kc%:ͽERaR8eZJ+˵U"Vh-+T"Vi#=NX8S^ѥhA%b0 i}-k][RR2&.w BU"V!D4M%bJ2QMA1iWXk*J"[ݽ.l~d/!>|Bi׮Ž^r#^jŽJvjqjqKr c>.Z܇-_X7XZܯFUZܯ4TjqS-nM7ZTjqS-nM7ZTjqS-nM7ZTjqS-nM7ZTjqS-nM7N|]fju"PrԁƊ@U!PE1u!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@u Ա\YUc9Fk=V =c9%:oq~@u!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@uh&P ԁ\kumuAPEu!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@u!P@uy|ٓV2:yRNgO0ʟci<ד~r/P{^cyPc^ywdͦEYY_.bR,"U- \"V;梟֮R5V?~>Q4  珘^ڻѡBG=+ir J"O6b& FJZΌѕuY--ݛaMR>e_F>_1P1?d߾D2V/~{gR.#-wNr-.jTk5#ңs뻜(UI/NӅ,R0r#6|bʚVäMmY{ﴼWkV7[I, h {/BX=* ]vkGyAJɸ= +N558Ŵߥ2zy.KTWgpOM-U~rfl.l~օ_u;^ܮV"֕ 'Ե/fN/XwZ']%bH+[ف+Z82k/RJ;\ܶN4o&Wa wS0BR(Žn{ ҮJ1*2WE՘+bZŎ\k.1WF[b_sܪjU}-ʔR;2Wo\Y͍R+RVc\WMtc7WEdޠ* 0S*\NbV** s͕"sBՓ duGvUT\-ޟ︡~>?bw?:yi)f_iWzj-fcQg/5P>LbF'9$b{jAY}ҵƑ&61"xJWa<`'|{+˼]w_tqW7agfR5o}f:ۗhnҾ>t>N`Ǽh>'/7m1g|oOo&og'kRm6էNrDNDž]F?O_h/hq=e!-s7,Θ7].w7eMY̮/{;zb{)C",HUk[6;R yu[^w>Iuۦ![kջuk%XP@ Hz9.;'悋j'vK|Bz\)}Q3Y~-y0gY'QŠb|<շ^|#.Mm-n7:3'/l`Hj90oThnkYLZ4a 9 8':]fxc!F2_<ݹzu~ %:rKd/"@֤94l/)9UcNA6v|CoWvags"τetfH)|QkBTk ,qvшzPoP8΀90#4ZZa#/8 s]nqٟ/ӁZ-Bάޕ+Lvw7 ChWy^H?Rkgl!.}Vw{VikAT\ᙎuܻ)/./.8?ryobU4V^Py%+B̿,H۔RN |]IG ϒv5v2+6Mǻ՗>Acg-c3U"`S"[pp?zm~jiQM9/?|}NI>ǶFsﮖv%q/ms1<2Z)'T"1^PF)ǡtYjN1XiC՜Sㄏlw,PWR!"gl1n>o|Ze][1_Eqr*#כ ra]^\?\X_wlq/ͯOS+Bî.M9wR}B9~zwo8P_Il}8ަ/~ۛ՝^_wMnj4RF[{z>H_:Zq1/]zfvZ9}0ۤ04X:_W{_ߟ_hc[^jFq,L^,mq.l77始=Y˪-Q*X|).2nWWWKC*ה]&1]_l☴]DT~i)Z #!`sbˀrރ ,OP;i@ ;aawXMNW;O=)-ai&8;( ybѭ2$UqF:Ot}T פgz!(syҺ-Lߧq۫mXߑX3C ύO|`xd@4P 0k"s`zy5ѓn!1QtȥhsRN`4$L]YH[z­놮"jfr='/vz&AQ̿v遝 }>_Cr)9'U󤨓3n/dG뿻bͮ&vnH2}UkY irP6+]oe(iDV/tFo/ U`) `iw/K!~*?:qfXQaC-$sA08`)""0EA9e>e,xHΙu R B,@FP4qPyTwn<]9! a}|U 4۵_SEOHxaypX$cV~mM5Gi;ښqyN??}Qr"e7* /"R)nTpXh}tGZKLvj/Z2x,HႉN)KN뾤s~n2A,^dM+JsڄäS,񈜉-AGk#]ᤴn\dNwhED yN nC5U*%)Uo|Ym>Lc8.4GSTr EH%qG/ pnuEW#֍?6oA$@.%SHi1JJU!G=EWO Jd%ZA>Švڬ%@@?KQ`sA,bI#t8eb U9kz9kyoޯ/xw*E.oxm"+ $.08:e2"ѝ:LN^FUd^dEsa8t:WZdL9Ђ ww"RxLTdtf9xC&2k?+fÝO®SC>SdP.-ޮu,Iw^+(vc][#a/=崡e^jsѱRa ypJOɪ'o(ukY+dל :a7M`>3J^rN0l0 9Wy~CytKF<7Wkdԇs2=C[R6- @(4}o޿y"~E3r8pVkTDJ dLw0X^rC &rOfc@0\ *SBStG[ˏ9\9};`NQ5WPg;,~4lPvUw"c5B`WwTsu^Ԅc8BB6je1´5".U(V(c ``b瀔ʰTju6,*Wp2}ON0:0}nku}.O"nsy?T?4LD?uRKDx?މww|[ZdL(9g*&X-!T{ Z6x.?5 p >z5guۇ.Gqevm1ͯr9L|uhP!/ 46-C򳘥=/M%R P4]%RGTαH hq?9,j'A^њaf Nu/gvoܖmpe]&d"im1& ./<6z|^.zI"I29^ l:8O޵<`&ߢn;UWΐƱslqS"ܹ'C2 xRQZk peZ UJry[؋A)r"W#Q&; N%$9 dX/bT9K{B/;xTz]ǰ]3B3c`~sgT&lѧ aOwٜ sv=ˏ؎X2hwR舭d7.zJ%T{rFz>ET'O %9\|fd|z*g)aה΢Nm9Yѝvɲ䡧7qmMb﵆Zk+yՍfN~-bcK׸T,dZŨDX- ci؂Y}??;f: P;7`ygY¦cozD Qo7J1C- |/ař*[Haƭn:,I>0;icwhC&7PM+>S6q;ER'wezwf;?`: (XU:Aʊoy ^zbTVP$k ?=zknV"~f1vkn^[҂ۚs'gO^& pYGsT+EhqGTGp'Uܕ"<ٴQˇ֚$䚩ν^922>-z_h̦ T&@S19'md,ᤕiRPC'`y GQpb r<41ˇ|DR,ϋOK`"ho:uO4fT2vZӃ8 Ty[Eq!R[ pSȖ FhĨE)JB106CDCyVm1)Ǝ6M.T_^pMfwx!Ѹ:Q =H߻M|` 1(֠кݭ"|1_; L|T@ b86)ueS pVX¨טBF-*d.8#cw*G |Iha_;[/$ K8SGɖL,(rDX |Pu; ܚ0biT4P8f(|tGL^{P]uXG Y)yYgqu|K9pq+k<`Cz W!-edq"6U&;GW1YQ) ?ΟWpz ^4jPr?O>n$6JºߏqL DПO=J)X s8Ul˝ؤcskھtIj 'r"/$L|)&ل%Z?=NmE5ZpS,8ixy@I?qd pH2醊Dm+ʋ ]<wZ4N j@+8V/U5(yK3gKt(;h92:x|ppUdY 5^ɆlqˇdUㅲ9Ăw[xՂnሩD$ѨXH!2#i!{ UjGX fR'`60Eles$" ` rNfn*0nCnQ [gӣr(u\aR kTETUT%-CKD1r2elmMQQ&|C[L\Z>PcB6~cD{.oLP^9~>h/W},2YYz`dG0jPZ6!pc̀3#Tƻ&3t~NdGa:|$㸟*FQ4yvL_~~<3DTtDj4NsDe+/ӕ)(Fbn^C| *I?62_WrNkls}һ `]"C\xI !xrN%n–Zdw3VJfIZ1ZvɛN>NW5 6%Ui!R~WTpQU#VP9S;mnߖ$ahhCW0 \~mċXRLQW҂%(+€xDt])nda0 rꖏ=vNp͙Y,<1RT_B&:iicr~}t&{AU#> K5z%{3e,VRH=1Eu,p;'%X&QeN_/ԥ1{Wkj0}= UF:J"?[R1|FCOtv|#KҤ%z8M\~եBo6vL?Қ;"wĥo;FQ<~u%dB^?Mn5u,R oXbk՟!rRm͡$̓a_m~R|gXQ QW=y)4gj pI;\q.=LdXgL 6뵍#^՝eԜ2rBԶ߃ "J[Ў|X5>+"a:4F*)㰻v}uϼ +04e}$|/;E 6c\@V;sUOptudM!26tOZht̝z{Xp}0AsF B$eAY^wM;gco>b[|Zz^5J*g_-$Y7l9m$~/﷐K:ߜN=}Iy`?rzB6Vpn./!٧o6}Őn/SJzs 24xN0]W^snJXîeQΉpZZ8=vgSs7aX} 2yXMNek ŏ_44 zh'zC^wBjjEZl:5;hF^burU$|"WS1žw0(ÄESbh3{C'X.J?e\5g{ F*bg_ HLh]#S" ?ʯwY*+;.*;:. @l;O4LA2ZEh{>O;O߬{ք֩2A:.pL{.*&+}VE>Ͳ ~ztrh5 Q2hPP:~FGz`&X"vl{fWd/M{ HJRXU,a@8ntcp]hvޜcv`blh##.d"YvqazF?Y3ƕۗBVTAZ zюqNsr'9>!Fv.2gڰZ޾"[@v~.Q^.C{i!yip1aZ0HY.ϤSN O!7HDg+3}+~dv2_a=zFc 8QWMRÑ et?k֌qe(rq0q(t$'LɜhG9C=c\9 D_T)/HpIb1\`CeWeɮ^W`PSV'lO-MLPWV.W?<3Nnuf3<6s g1qy>:{eH@`otxFyWf|y[u.S`qXa1R{0NCOuUYf _piu>5uyE#eieu$/QK&+ |TEb$BzƸ2s3_/iqe\ ?Vg?OCX Yʔլ Y-WΆH;x"Bc*,8ɨrX% 3rgX grRg9̇*L N%OwT`M)prCX+lTn+3!4 a("&pt5׊M1Nl>iyVJ3 ^.Bl qkL]H39[_jGϋ%#? w?sxa:+G|> gwP6 u_ݍ*_W,&cT)?4-`9pT(l)Hc-s9m5 Qf_]e{o^{={zgEܙ0~X~Q۝M]못8}?ROJh)ƌJ1 8Vj2?%Ǝ1Ef btm1#u3ObCxܨc@CȮn{#>04j ! (M?񖳆^u!b prz>}K;,蜕2)\ L׍3┕w0DSSd?B5nR)o+(>9uC'VSU/1ZgJ> 'vTwԦš]'a# oG$Mzb[*,`Z0xua _ Yv)Gt!:lQƌ׊2 86l"l3>pT=YM8qnhl#ʘZQFɍ*W8^FFp5~#^#1zY1^*0AcE sWF,M(FK^[ڈ:K+њA$S&Nr|֥'  p k2(1U ^83BGK,W*ꅋRlqƌ׊3 8V5k2PMAA{ܼ*O7|`䎒x,#0Sb5Ω< !S~hjDRI#&t%"ڭ0{ <>k7|a".02)Yumk|k >1^w#8֥¥+EÚC0R! ~ȳot}̻ Z   C{ׯ)ZO&c=V$Q {)c1DN=~NPݻ>Vh x-*XBĖK&C^P̣[w@{BhլiƌJ3 8V5k2PMq63Ua=p Yة]X`TG( Cl!cSЍckEv5C'dJb(t8߯"ʸwrA|VyxxQ ҚӘ7LfzATP*ȅᅇP+`u'R;[)|^v p$k/k_]n>aA4IhޢZڟU`lU& A)M4K莾t {P5d[O1,/'Jmi;>zKX&ou^i%*LxdV4%xE>B9|yjux1^džg5HTg6ty c'hɭ+:uv<5O~tnD31C0%u<ۅIrù0YpbFrBJXnRԅ"_μK+sRD5F5bF58vb bzt,"nj1 vs2^PIPaC* Ǯa~#dyiQz ]6 s-~s㣅 zku-Td *)I_1!ޝ 0"o$1 R;s[H~n\Ҝ$k}ae=邃1T#-X-6ثv#v­k=#j=c{0T݉xZmAj9"agSjAȌ0 T#Y&>f`y{ɟP\|}-:aqgcJ>>pe[S6^I^_Fw4 VsgJV>d7w>4x寲뻢47eFpĄAUY]KᲬNd; ɬmF՛~֘;Lna"(^*?};3goܧJ(bYS@h&Tv#~ &8T_|&fYY~\c&֏/tY>?( Q~=W7}dۇg3_z;,+P T ˧j|jUT/-YΦf˟fe;aDեO2Ro+)X-hp݃s(}d({ TDjYeɌ@hӲ Y)љ ih"o[rAZҿFqj6M.|xW5[ ^vCf'o*־DU WH/>g㚾/!٘=͖/r-XG 2B9׽;o]ĺ"b5mS8l^V8k,mEy,PfkIX?=haШ>9]@Z_wB Khws|yse_ a1~ڏcGͼ7[ ?x^MtR?/*Usbj? zyiF 0h T+=6Qzҳ2mv\Pw,OM!V{Bڡj~7DCQW@b([g^f v"W"꩙wf9ptݻQf2 wz4}Uq@0X~Żяa@T廇B=;-Yϝʯ ^DQG۠FS:Z7 {Q]h/-r|fAAKQhsdoKṾ7jyxݦy pfPMA\.iKmm%8}tMcdQl&7j&,**^5td `4h*)xCIE&8zHnMIIMc38RDӋ# 4781UNԯ++׈1ѝ++2TSFgLP] 5G9eY2Q ^3Vxm^CI/_J|=Uw{} ofs9ze|i*>\}Pgf^}ѻ̆)7}tVjC#S"Gx;8{4S(yz"]IfL2aJdwVOnLqC' ; ըڜQ+BSUf5(*O3}E#~ɨ˝uͅI9?fϪ[.ԣqw?O#d]ȝ1h2`Νwbz?Rܲk>6'Gcx.cb"b#o#OTǤ~e)N'J[J/+kFI\ V7iF]l|9pb"䗄1@PBAy0"<:G)QĄr{,ro>%˰w)ƨo< }3nD# D`x"A4l/>ͫjfuM˞g#-ܯ$NIYSjUd `NareLAdN24muF p#ŘZ)F" LYÄȸH LW G͸:ɗ/hZC؈:?68ִd `Z4O'(ec!sDkyA$ʒ5# "9\}%s,,%F"_^}d( oo`ȟ/|PݟnDTzAs{$WLQ{wgZ$/W&cp$nj;ӵ*82f%&.X,ത7 id ^&}fca=ׯzܨ9tx| U= JZp&D^/f< ŝ\-)acj{_|1-=pA|k[,)j3 nZZeKlDjbUŪ&VI!*zR4O$(%x GV}e_ˊf"&ݯ)թ6@~3B9);̦vb,³u+4=a"g4ڒtd+ӧAsb(0#e!:_gK &>|Ůn k RSŮF/bs]4WCPh劃sq A !gn{YJmgLj$ hϴ;/%Uv|}vullE4 rh2( Hhk^v{B|_Yu_n+)0Or$3fQ b '0n}!`)ڛi)-z B.d@v y8 w<,C ~g 8a@-j%4BeP hq8brґ ! ;HJڲ;Rb1gvg˙|7f[~ûA})s;p᳟ "`KWyEE$?{? `R$U׍o&+ `_,eI`{o'(hhtQUW߅o"^&'߲c $a~x WeegBx#%XπQR\>#*ԍnjf~TcFƲX-zzy۷36qcɃY#ˎZ?K.FRTAYU#qYU qtpXtu[jjQ(|Q5 _B? ksnjߺ,H_VpN[ w=i)~[I̢.ߧHlv]UUA*Zt0uPGi6.7߽['>5:mP5~Vb l%'OP+-ÅwXT4՟oa=o`Pk.؛S`/b!g;W?Gs*^rI"NΕ˖Cp.դW_TPsBΥF-g%r'jne}xS10t(%Dsl5b*m\[/%^  |`֊_;1<PZx V02*YPv/]Z&R0iP0v7/V°RzICwD3< '(Z (@9{̾JiK{o=GЋ[^\חI=Ů"suK0)si,*>tRQ95c郵.v?u<Xru0ʅ'/˗سQ.(^U",`hoA¼OX% Rqy0FhUMboxfvP7]rJ?jHi:e5*y.X/Q>03nգQ.bwK^ww_/Dlc+M|Z ,a*AMho1426䂎6%mѴG|^&R`.J(Vx@UQcR`ŧVM"4)0z(#K/q}rN&skFE(gʽ{1( VX 4JarWM""x?`Q #1-)/ |sVqfL9鲳m}}2FHHܐ (H.0y ȓDI N!>($+5B7*DN|Sbar4 <֞`>/5G1)p Yc@9G t@4Z:T3;&'r#B\2t8Sp^?q8[}̉ zT>y9R9ZR1 tA;Jiqɬ$3b5##3؇iJ5Ǘ pb~F"rЇfF d§;ZS7ńZb qrݻj}x%2JJ(ԅh-B+X}Q)-)g?JթG0cP 5`IЄ4"|nG>q8葟N:M1M7D` hrg _g.{zVaET>tJe/5Ì B\8M+T 8ȸAav28Ḑ&!ޮ2rM{Q e!;PpqP cpܽ*6!81nRl#gC9Eęq-w2&aLR;)QSg=>N;tA Θ0c8&)#6Z:OW yڑ ȕ qTp@@;Bcr8*HiS Z7fF(O`!(hny vz?JOZdj}hA-cm+5*l\AQ((!68>d5[B5[ q*Ga6+$6 R8X&4LP!_PYP,4^rﴃk w@|v0df4N.~µ]0)B`xf.T_tSG󱠑^vd 2&2RGCଯנs7p{<2Z)'U!Ҩ ÿQq(qBrAy4T=ƺa Q=[ 2/1,g*iID,t~<x|ىL˧#@C³? Tyq8LW!VVgB; ̅#jQ{BvíDot$ì^l@ZwL?VH%X/,GP+7R1r+UrU|mW+zfERNwBw#=!kPL:9f!D;IS iD!P? haRL%zROac 3^|-0@EǒS(($D譹-)} )~lG \<}_^K~즓x^~#w󭭳||}U7\nq jf1K(xʺ5̩cbzH_jJNc gpoXSh3XqX4jj0?g մ:# 1hj/}8wHr Иk; EA}bb\,c;yFͰ5rifq9OQ#l2]9.j-kd*2>L Jl~tu̠v+M:z+Z#Y>ΆyC aXj,bĥ])yEn>~xVwhq[drO/iNΆ >3'cƎ ZP_ށ0Xqi»h12eD<G5êqU:Ez -w7@rp:=ѿ\([WQ _#.-^Lwٯ6/b_ɔH_K"f^1b3a%ӣ[֜lga|iMXq2~􀏁xAw{bt8T"/N-<[փz0/ M/?zDp'X_ q-?)4P(S0 , &D1mD;F(;J,]N/3_A>] B0DۊaiL6;y  d02?>MǕ ۃPqOm`oߤd4յˢI ]Eg7)8=SגzNnjs" y+|7LhWM[Yg_w%M}z7 %uO§e޳b:feq->6j04K[׬ me[D  ݫ>%4~66M'=nn|HZ1NCmΗ|9E}¿g:6KGly##Ws m]*ތ6o_cmIp ҢݫArBLˌg3_<'};٫](0!PR"H,H1%:(˩n4eH)-Ux%'_'HiQ8Yൈz@J#h#!f<>Zب]EGQl5YEntPigQLpD 8 ʧ%Ե'9\R_+*M 6V9{8ݽQnfJR=<~t*˷IncS*E$9NlgJv2Y  ;jYsurX\ ҿ˃GW`c5\h9vh6D@oi릻A z>4^_c&(rPe-砮8FJ>ev,c۷Q2*b{DXW[bpGaHNR4$X-73Ϫ5H[Ea9 (QcI5V=$~-.Tan7W-B"QwaA8a@2׫@'GES-_LK{ no4A[\t+8$OhP;Iys?SHֆߤ%G5[vr:G"MZzI;`1 qȤ9ep@ggA/QțU޿Q|܉/w@M%֟ k _4#lcEb7,ZeYM~zxGe/4EY>eՕiOs?Yo-S 9ҝ8i4P~*:W} Bt ŀPdĩN;aMk-<&孂XrP/yMNJ\UXǴrݻA a^216^$ ;}lhoWn^80"fI" ^IN05Dȡ=x 7Ϸ6'G,AXoUS!24@ :9!pP$UFU&q9 U,QQf%]b,QI\fA\PZNZ2%@HXA+%DbFA|װ6Σ I qIk&*cV]>`խoe< CK 8+aTҤ,h 5 ˊXS BYI\T!eMbi7CxWx)L`Ls@?cg)*:fl>Gާ~BJ4%11'1rN%4EHCrfHr`3V<|ԭ#ҍrIh,C_V+_e&sHD :js@9[ߙهUiN9 xccΆP =n -eD9#xehBrXwDGV%- YbX$ χq5NV"ӱG1|#:H07r8XZ_: \Ј>1#0@BD 9niӗUC}qnD~Ad{~cP% {+s`lc!(ҍo1pNahf4A" q&x@- (v<mrY"膪 *Sw8FaO6ފL)C;CL=5TTɆ+C|Eփ<ɁID3LHʣKvx|?W7&kJˮkFN"8\IL =z$ EP CeG aY_k?6pew8Öڂ+d| oIGcp*8j]݈h_BfU`(PhrU"R=<*Q 05ē( c9tk$\7$` F11>FJ 3123.|yح BVnހ` S`XORHT 7z 'yjBa<ڂ#*dӐ¦rS9Ft4Ti,y<:%~x91СEܘ:塢8 ~cPxa:[KTNch;'ߍ?jV3KRIGcp଻y|BO=j~f}Ndɸ.C5C b܎Eo=?GU KWȻLDCtA6'9JF!p/_"~_4Rz{x4AB= ``l/3qfˣ18AQ (|h LkoQζ$2h5P)Z/M|TX v1m['meA|2 ֏+geՋF/RS= #(xmgTǞr8ly4GyxNv>>_ncd##CI/hG hqod\^N}3c*EL8'AuJ%J bOTG}n\|Q=ǡ16L^>ri +fE+ve**^"K+#Bz"'.Х]M!%1i [ P@@]M )j;٣U3r@kGkplU4I9Fm KwaQG@L` Hhs񂢄:K))v]=@eat62)/)[?U!D3`mIã-8Ҧ?:cay/HZ4|c5 k*.iu|M#z7 )]Rjw/+ee|cH*gCGQHf8Y|: !P+;$ +WyMil~´#KX jJ- 2:NY5Ci!(c2/Q,cV5 i,/;1DAhm&w{OF"ɰG5rxĸx;6"2 W!,k rKzx4GXVhDn:DejO6Q6eC4ƨA@?4>h XV%1o)qPL'F;n %ree;HɷJ<Sa~o⮑c>q / X& u,*^J4[U6,kNcR\"$4.cإF{?j@tWGʩ#frS8t¤1B|}rnLam5Ek)7U?ގ[E@¢M(6Hx9\h [FL;u<@F:mĀh}=O7+iQjQ秎ܓL&z\7{}Lã>ER*Z6ã18r5)Yk?dUEOFA^;g8Rpk$e6F̈HJ.CƣoI,Q %#!hbTMI+dψmRY?^yScaɎQ??X4lã18}@s5]ã18~MP4*^*z"  tSBpA-qMVg\{x39(ã18ֆV(h VaBm]g[VXmfІZztH#P0oV:cL$f$DG_( {xgG$<+kL5iw?ѣBV;OzML2SM7<װ YO:k9-OY{t?[,ٶo)jIړw/fwh N7lBUVhGdQbT=QK)(UIa]&>C*'7FƈvI1uڴ<(xO1h rf|~VTh9Π=<'|ۈ%G7`B]g[:$-OܡZt5j9?] <%%(b5&/bW>Š1,"fFg !.fGW3~rb}GcpQQYΧ !MM.tvE Rx~*df؆;b1_j˪1Z!UK3H蝀y eFݶ xK!?6 )Y"ނwFD'`)Gcp*,vxԴ%ÀEƌs$ (N KU&_:&65_j¿wZG)q;_`83xG+t~_ԧHb =z {1#zUĠo<^~ڗH:%,:RvOfsr[0ڔfMׁc)^Q+Max3-(u46l*vޱX_8-b -LrI'ؐZ -vn-U kyeX7yɪ6?= W8E/fvĶ.C g(pv߾e<8vj;P῿+#'Nⴄ='$|uuu?+iu8MsEMbE qlr0;.@K~O1#70p$=+ͮ7 tknH' װĶhLҜD $݀3ijtcTlڍA ùQbIDY0XބB`G)/edqJo$,-EY)Ff˒$Ws9VԦ Mr맴iOAbZ_SsXT)cC JNU"S&3}_819qGM 455 B?S5ʭ2PՍL q}0 (UGjyT(bp1HꓸA.0y>ǘk )[k!7ڌnOcmd$P2YĪd:GdII +f*_[ًY<3zuv&073/R r[人ލEMSзXl[7 m\#w5:iKHN,S[h/n)di M##Z,Dj=9<M!fBa,[]K>J ?JJx}<.31D2lJ*̅cJDX"ؠX[lhU>^WEښB;vQ `#r@!Րd,U)`T'm"?|w~fbj-a_;r?yqMZPUƏVCenZ`?|V j&N;hځ-&7gpfFǒbZffAT4K6va(rkOLAi&ry OB/V-˔bx 74͓lbY|.} cjĊ*M.ɝ{=,hLJ~zU!lhk  춘]M7V-=3tnwr4Z/V7k[V+|e1VِxrR"2~*㧲鞪}Ս>c r- Ik5ra*Z+\Jt"#X~ IFېx$)iDv$`EU=t *;=3i bI BJizb猘W"h y n|ATV"3 'Ѧ4T3[l*6H77S5x".DSnwpޢ!dЍO22,C&=Z{\|~\l؅To]H[`Y*gIJ@.9-Sβ .IY^Qm}sm)WESZ> `@(?j;{h  Kgaz,LdދZ݀ D)#yt8݃X6Z~ $0w^pQ ^gaQxf'qsAA.$LKs1J;/x$ek$|47y<F}$ teeZR1S.=^HT-R~_yDzx?t%3-9|LBrZ`LcUQi%%OCrmHun#K 6Quj]Ыآjj1+2a״چJ4 3re*k.#9xUS>#ɑ0ϔiH*:1UJgytyYI>uH.HU*'f-f /Gr523?Ax!~f R\4压SKn۾qC7N0ȨK0NvZ}?^ 6jBم?O\N"3PZ{Uuӗf+af2OvWm|[ĪUC'*𴣘¯ X6s˖s;g;ѫQBy6qGzy;|बjG1E)jiwrd?ZTZ`Z&"~ QIQD1s{ J2*fXZO]DçKOp(:1si:Hzݮ+=p=DtG?tƫ/d4-91Ef"TtS [zSXVdk4/\p^H>~2|b7(wyZs ʪ촎w =ϧ7wk!\MG0fk|MZɌVܨԲ̐0 !YƛSN0܋xBV;@DMPn|,W͜"xGF,"geb+'3&~QV$^ʀ''"۶h)ʒqFa&mAhe^Ӱ 6 A]kT }q܈a\ [k MKL]SĜ6beˆjl :V0uHg5sG!Kq VF*u^}Tk\0R)O@vt=bh4K9ba((o^P[ |ekAxtJ¡mxFaumynt _է nvU^We=93 Xol_g;BXڸ#׀T|q^sMFN#]Eu) eq Tae+D"?fIqf2Q^ݙ%ij*oHo" S-r,fu4LV4)'mz \x170h,z!Yb*(RmM'9ަC=M6mU G1_=2ݳ0|eD e,mXPS GwӔ-5c[;wȟǜC-̄ # ~_\` {_DG* 09暪xe{f36<9k24SJak;90=;ߛ[M{߆W+.M,5.r~,VEr.n){搚^ԇ8u\!}H̖!q Kϭi罅CD2>8] F4)jUd~Y _#tVҦ>D΁? wb<"aٝb(p E Y¦\Ш7zHQ|ٝ%԰Tf: 5ֺQQ;ȁ~wh^EYgwZ}2H9MV F":' .oz?u~8j$4@ys]Ze4&7s%q}7߃UZA81"U> ГŸ,)8{vЂ=C0B's{';ѭZZ:vJl}x\}pnUPߗ::-ΝՁYΏ>[)Cx\ԛ᭷H[%pwTYNV KO?;~;OE"VVxtgW]`Âǿ?|XAO~lƗI?A̿^d/h7P H|N[B f_QA[YVGgGǯ/v/V1{L(}x,bc1'nFig<%F;ܴvSy<%4ܞ7d+D_)x^}H N 't=tH"+F> }q'vTYW1+?-{]ej*T~WW=pdD>_-MnlGy8Ka3fO?N'';r*O!3vN"l›?-X<6\S^G8(֙j߇!H?^t{ uBdٰ?W 4Po~5n٠!<:.Pq=7Y1OrqE=|&r/xd'f#YaI#XrJ$CU1[w!Evl 5HZMvX(BC$E|#E BO:Id 7s}r'Yy L6<Б.kt eewdLm("riU@45H h#C172ydžTlW2/K(ʼ2O# \zĭ`Y8@Fɵ?lẏV}owe?V@2nHLiauĔMč!kŏ94'~ʟw}xB'Oy/#i`,|'0U`E}>y?Ab*\)D!Dej>) BbXيm6%Q(Iy=V)+!hND탓Q$n_d=hr톝jQ$vhۇO{g1ɭ9>zq&.n{kisTBs#r&404ba`8>"E~jP!,BJ{`gH>,G~J3Fk?[<[?xql(X:A$޸NֵPTC ugQ!% Ϸgg^,[D*JSY^љw|~٫c t 1@"h^ +8T]% oA)|KQ}}w{?j1Uwpii8:G'W^>ҽzm3DD*!)rfZ  (DQf~kȯC_R_,/Q1Brl9P9qF׊gt ‰ F a(5AE` {Q @%Xj (ønz*זfz368>HdK_D  [uJ ,~V! aVIaV)9vf[ ƌGTmuRPCm䴩5u~)-D#0oЫhMJIo{WG)FM&0`!nr Dg *4m i*/w={Vts>EA$;k΃,:{Y 5U=@aV}2[6ml-ZvֺN6>|9|9KrN6^85[(9c`6JLXyA|z3R,1up?3K|Gn$+@ ;/uRՎ*mv^;8l]xE\%w8'^N __F }}sTT&A8 }0?)4d^;vT;v^A Jm+A_7aC3OSa}uI2I *y$h 0Ap PK/A@!Pg:(QrK̒+xe_Zg}'ωb89I__q ,ݑIނ}˻Ƴeywo@7ɱ0,tpj]}6W SEvMWgv%lvf}ٴ`IJ̯)nޓ֏?ii{i/&O꽩 ܴor,;vv?{cU4GҲ[nRAdT;n]7B2R< 7c}#Ƿa֛>/sӷen.j2geB_Vo "wTM>XDBEMaxi1d2olm{_4R^. +35WYw#7  c#j<ћ"e0yO2m~Q&Mn iJQĨՖͰgj,ڜH#D#WNDԈ6j[πsNKI9$3}ZD;r? ڜΒӰPSG\Ɉ 7RB2\:jGv^ÄoJ(⹾I/hɍKm33j 17od)NFE "f jSH/?Ϥ囖vU{&fE~2!ycCl Ƴ44a+7ǛVtP_~o I7%pS,I"s}~dG x>(J1Zs%1M(Ia-I+?\Z{H bpR3a30+4J¢(6^_S!qNNIC|BDֹ)(t226M$zt66ӛnͧe²%_/5C]/gI7qDZ3n־مn3ԯ?sfQ - .ɟ??C۔:*~\D7KM~YaCVg3~i`'?Z^uؼfk\aھOV&Sh(+w<:hj`ZWҚuF0y0J<9& n®& n®& n®& n®& n®1r, q䧫\sNY~OC~oTf~V9#É}=Nr1N*{_bj}NMFdH|'7Aʎe!Ur6|/1"n?C\)iGݝ>iYMsE o[G!@rJ ‘$(!Q0fJP a,#2]!,ZOiQmV(NeSTƩX2I܀Hb.WbnHp@| )q]W "ڍC|un'[(3# @ `Y%(ב iEf>ƔE֟ up#vWpW 3Gj+ !#ȐČ0B.Cv}jpVS {)0B43XFc0D)`@ aB)bp('Pdizq&#ʸA1ayDŽP!:B1` :m:JTN:G(47 L gBC# 3a(ɉ>C XZL{]VYKb?|6`A<f*WS1Z{cFrׇ'y*{D 4#J5C~̋e_Hq `Ŋ )0, (Q( ҧΑPe*{/ۃ~(?vN69 .a>h[ saEp* oXR?tb:G/Gd5t\ycwzNIձt^L D*!)rfZ  (DQf~kȯC_T.gg )``+pA`qQPkT>`(t,K4Pq;xz'H?dS), Ā[ %@jbUvDH0+XI߾cM"vb8v|>K5\wZ[ҴlП^˷|+ A}f%oRRVbUer ֝s>Xw:X&SLD*(l,JhE4)< J17A\SGrFfa;L\,I]+ AiXUuyVm7IuXM9q`VړTNY$y +wǀ{0&xZ.w<|Mg%b69ڿ`O`2=33CƊ5*rcCkȹ/臣F az;9A &]l'jPq+&]WaFv &޻qZ d@L=r󱗼m'ɘlIjgI\c}N[߽wg;?X*'V v6'>d`ھ6sf5Y~\{^AUXd60>0Kg8qwn&z2CU܁ 8ZˎArhS֛kl;8"aq~;[7ރz{F}MӗIP͛7Dpu0a+uKv^~*ŕ _m歂(۾1GL egVZGmCXÆ;OzOگ'DNHHݝNYG6dbX{A+aXyvQoIIaKt½Nh!χZx o[x6u|Πpk<-뭩Lܯ"Oޓhq,+ #f_HH.cp@>9~:=3= $E#a_UlZ=F9 ꭪^Wx_ln:R+ Wy^ iLO]LtZ\-Wi~/3 .6elʙ'`\yg,jS 6̓Slj19[mP@Q?2zlPA;"d&3F<xP2|>( auhjq:`酾9~?r.q mvp˫Q$ 8Yݨ)rr;6O3@mC{ͻ_|OxF㣽d-:hkk@whbnD96asn]?2'. 1Th%DR;+#([e +.mHXy wbYa{0Z}hu }Š [~wk( Wrg.Iک,(1,FRxWRU mk 2fUhPMi9È~$ڻI]p<} ¿Z?:|_ XޫdvlxVC i f]mE)yh:w vLBHY!"-O݅G[8O1iQI W)>E/H杻 J+07jve#ވӑeg0j^6>Wp8x|* ϰH^;4‘߈v{q [ @+i ȃv$ՁUBǂ ·V ;^yUno`췂a t+v-"Q<ijX5h]Qwq<:LM<47~l'-_:^c^jd6Ձ[n6WHޜ鐒A{%\֜ ~Gqaq-z¦  pnf"ݶn۩pb㩎5W"G()mr@7gnð-{O\ hj`JM :d6I"_+!ٔ0@;kNB4Y?!3~۬^eGg kiAA -Uٍ) 8< ; (\Vxyl89:h$"{ 9ӀDGqu@ avX"Ƨuٸܒ @~"%y)I7^poO<-A xy#޽^vaȌƍx[!iT5R!y&1L /n zHHIQJi>rblCSgfL2e>SOBP*?QtDJ%"+a}#Yc+3Rb#4)_lZV})9S{!!l(8UY/=MueA/q/>NvXouV\Tg\j{/6i21+rX+r|pWpE/|:^~|{=ZSpx^耆mJp'tx^*x}קJ)8Y+cnvu0:~~?)~ij7Z_vۑJ?l1U(^+VoIe9eZS~Ğݬ_lpr`Ň6bi^_WyE{+*aMiu«n^n8GMz*\mSCC*iÕ]:£ŤWot9-X Rʮ6 c!qrAZI yH#x̗>ń+kJ%eQ}W/GC>i$%߽yo~mQgDQG6npEyey-'Vs<VL$J%<?oIN+YE*Lm2d/ ==\:v4'XpUx8{(e9%;iSi"iw:˿ImB>FǥM1(;n&CwCV⸿:s[34rSC]ZȝWg|.Ho7y뛼Y,%af =p! LB ^"l\ggznkЧ٪kD83qgTuѭ@ ,ͥv`\_w L4l3G~die;*\sUO{}AO:fҎnX$[8' { 2[nLB-5͋1 v_Dɻ; _};Nd<1KQ#͹њ:QFwt-jk9azHYFIښmPML΅־aM]3RW6"[IEIJN_AO^|@Ԯp5>e# ,ۡZ b* l`0)dj`LwZe,TꇑH̡t6bbYؚƾ#\MUmk<044/SMl1 cPG3y֔ERڣBԫaLEݩU= gpųpv'ǒ+&Ep q uJ=ec}G^ :UbwV19 >VXAc{} \a3Fd*Œg'ɟQJr=M˫{&.²h6o_e7LSG*Ě ZžRنxO- U1}L9#O,ue30"/ 0CG1]W<+>ytv ` G6׎Kq~ˉT-{aJwNNA'ԡptJ?4B=w*  "9 Ǡ@kZckE4k3r&-h^v@ 9Df\h9x0R:J~1s<wNʏk_͚[@tisjQ% {[ :wsѦ /S$ N/h_O"g3yT0<+."i~19[mP (P1uh8."ZgR 0V&=nOݶnUB؞Y 'ȵg}[e!/5G|  HT˰l1s#vT*=YΥ,v)˅K9IaQeNx^djvdFiZy;*<&j1Q:ӘcQoOzv9_?=x]=ch9e-d,oOj'w`Lm8E:x}|<^:J(4w_On+˩3BI]t]ڽ&XDA5tfH"cB{n×,ވM%"CR Kj>Vxqtn+ND)ɸG~a^UުTNtڪ+uNI($ 0 stJM 5v[bCIֹB_*@A:mXk@M ]?MawDϳ'URy4|'!4 ZP5-$_X269yqxrXpvO#lЧM Z ldv(WNnzB"w^<D]o@&s q2nxa݃o]V,#ػ]R^I,KiBJ~!?8s%J2W.Uoi/`}`粟b[L(Ng䣸U$﯌8g~r|b:G 8ex=P4HE/A΢~]ˍ*W Nf(XpTˀj:[.3,6)NJ8G(dO>)*wIr JAZn\s 3ÚVJ$erA)74Z} |\ p'Ba7㻍yx# 9C&Nɵf鰻d"Qc9|}^qLh9vHr*)ՕS3#8ܖx F2*߆Z4n.;,s̗Yn@HOe vp{@@hm5* m]<^lzI6h˥! jwa vK䵍A$ |-!sp}7gQߕ 9vToq[#QJĄvӤ( oWh[ T^dH̵VuͩUM|X6狨h &T uUWU. Ե #Z7\J|.Ѓ= Hn?;a;[GL:n{pk%bw]ݜud$퀯;Sa\ C;.BK0&>&"W_[oH `X]FZN 0)X>/74v90qVfXut::{aVO/`Kp^e}=z8>~LB*ӲfW4 -#n='| ;g6Bw9+g }i _訑%pfjE)<䐜;3a+S_FBj+ӣT=|^{كg&+8F %*WDvH;|5]w.8[o)oo[/O$,^\Io)NL GcZaOry<ǜIRYTܣrbwLBNkpjLڼ?A5Yҗ {TO§3x4-ٟ4,I4>־68}ο!wKhR2?jWkWi\X /4"cF ,K c3WL?$Zw^^A 5Q\K=6s,^B8wƦRh$NмzY͸Zmڭ% V-huq<6=%ԛO?0s,Mԛg7k[dYAt-ߏ[V2 -ߦοZѣ@9^?0`Q/%٢Q1b~;\ЫoXQ)dy1mWn'MzK6B0dG޶ tnpWIr--GKצu5-W*WdTmx}zG;kl8p6i#EVvY7s찏L\X3Au=XfmCʸ9.o@Y,oh*9Љ.ID_OE_Hli\@95_-t6ra Ӭz?͖;jб#~sg~I3ceRNQgͥp9O&VwdVwQ'+bUUa_qhxȢ@/rXy&)ܧZWSn#)A37 DNGwn1Фt8r b4uHe!+%\<([ɒQ"aᒜ]Lّ͹֕FӊxƔN\ ]ąB>El>}i… 0 IK MddKVwz{uy~ݽe4h z= eYV ?͡R{UX\G䠼6וjq\BmpǓ6 zXZ咉)2;"#P&pu}0nKߝ%?V4?1Ws:Kڕ֪Mmu+=3}`UP{K-BU!Qnu$!3J3r,b;o3f6kߓ$,㋰P_@o$s9Nzf8 -]%_k4^go8'n!^{/xNd*{gYmŒyL9hEFdξ`"}M{wu-,4S$U`+y.)/5zf@)s*%pg, 9^Ǝ{⇒]8kG;? Nm[nPީ}6ndŧӇ%S=.@,w.7& A-o8*=II |*D9,C{lVY,96rqL3ML IꝠy7,nACs7z}~rX >\%:Yv\A&y(Ai;(%$3PZ}c}_L3;&|^{كg&+8Fáͨ[%{F^߅ػ]Iv-'m!v꥚? !u3RYx x6 7ߝ)h) ^ՀPƃ}2ႍsm֩nV'w5iyոn87[*0u\Uʫ/KT7ϼ+?4h^Ts6i6%]1Wz[ᄑkvKmUJwN?% ɗ]1m1Y4{ëgL=oHT?.P+2ALRqTQ1t c|*eF+y{flaf{|aw{(vn]{o]ҬkVqޡ[]d3G9>nt=oZuخlvgn9>:Dq-ηa#F+D8hmId޺8o]R>Fj BRL.;&$\v 6WYm;om8=t fnf_9O"_6\K~HO6 D9݅m>j(c+o^tȾLUm[!nx >|1q0210s FSTRͽ"Ls>l<+G &cIq;^6|;[OڠLttANn"9Hú`C'ޢfEzPέtJJ$W W)κ7޵77qd21TH6n˂K- nQT?,)l^wg$[~ a>>%&l̬V3fT2vL6%za=k*oiH4RҿZ|yZt!QF-fhvZFPI-j]բXjHFrrn0>ƣcKg^EӷbER#-CQ$ThDotij,לAz\K˰(!p$ LA)FQ), ;$1xFZ"@_W9eUa;8~Э8Wng[{eCbpVWeϲz6ql[eGMT5!'JPK?j*`Ks0|fr0|斩`̥WkG['cn[X~ĀhdMd cB̰N>jl߾}IWwwhVGg`$1,?PsYkkg>Ql t}Xb`cN-1yM164N' mȿgPmDsziB̮0%Tjj%*Tjwz%_YacdW%ɭk2qIYr\5^>CO0cjZ"pg(ѓj<ԟs iATyk Xib$2`dH$ĴJ:Uu" 6PXv`Q9bQQ!.(GNp$d(3| % P `(7aqZYguToPY(1slN<"]2pwP4,!i(\`em)!j@1i)+6 3aI|| %kByi.BBX9%5{ Ҵ@xψ.>HBhĸ # GDwPԦ0x\'$R[͈GVp"cP(,+i%8xld=M ܴU| %RS )BBRk%48xׂeTwP</"ቈ7 3IZt ").hI\FlP(R )#g91[f!'!J!"9^ZA]q"-.E[7 KQZwPL`1CSFFNDBᒶ](/qq\J&A\7 i +C&A| EKZz2#)X(d?#QdYBx%yyq3Ty, {S(^ R o`&%rB <}i0g0d(5XAt"Eqk/(8B\)gZ8y*69qKl߅BIFpWgം,`5r28?߅BƔvuBY)8G11\k%ՐhLH J'pZ'e04< ! 1);Z ^ %Y)='R%iIH^o 9L2<%BxK<CgD"9 0d#-a |;"yxX9ȫW52FQĬE0mP(aKGE'E,MB + +4ڷLtP*h|1# 'lTx$g? uh^!"- guهwP<ڥ@#6-n#oH \Т](Y+䭜4F XEO$7P6\yD'-b߅BK6_M֙]%.&k0o\RuKd4aRˈ" d  E>p{)5mڔq b[q(XM ΫY. JIFs?*FY$+TH0Bx*."1:E\fG!Nm{P(sS|xeKyG_My`i[Bx!G.Pf( eOq'I{)[-BxL%c C:IKu$T %kK,F))Hi^ONzl hɺP(^8dZlԐI 9rA&DLQ|wPa4$8;!ijT3Bx[BhDBC瑌1g]ZwP|}<܏S;)n isM)bI! dB[ y8$xᗻvn' 1p28$$!EqR{oL,yhŒ8aNֿ#"͵u_ׂ<х-)CҝJ٩ #D=)*jGsq"%A҅ڣ dhG JʼݿVɃ7N{$|ɘDZBO`;u+,y QFpY$r7ƪ&Q(IH>)C|ׇ=a^ p-{^_Sex!l>pц y֣{Տ5-rc#v2(Xjy?ˇ6hU㽞B ;w|*ql4jY"W@߄riMPd֏>9ds_{SxaZ ŁlbQ7[e+Pqj8)C(4iL8qY;8?ܿ}xz#Cydy? G|\37'qo>]agmL79xkw[} S$6eigOO{i]t`w0?`6cLwv]^/뿬V[\(٫/ ,-jy=\V<=W_Y8O: _^OcTrãd(ώ ArW߃pq.W- Lͷs&1mA74h{plO9dYX̚94#0`O޽xi3kZgCg˧G:] Xi;;Pjo$Wcz}u_7dytoc$fU9ˏH$& H 0c=2&gƀ.)c9&bO1ZkA,f}ǵ;dTW$f=??1ՍnVIgڂwxh8V*^ʕ-+|z㭟s~߭'/Ns=vVho?غTeAǧܜN>Egݿ}zu.hZ!{d~lx'ܿ}I8V=/ceyܝv!k]aNߘMOn/ #Jh .8?烵p־WODwju]ZqHO1Ưv6F" 4Vv5 &>T,"<}= Ry~ezTEV^?R7iBկ?`r-\o La\ x翲˫\̵zzqާ~xf]xj\ ТyLr8Eݛ鬾#07-dsهo yX"F=mlN^e64;ة;_~7n}s# ^Th^ݮ[pn=l'3elܼ΃v:X4 گ}s y0R}~뺽?)ַʺ2=~+p}0>fc&o{SLճg@kTLT "\} ]DӲd}B^s}ZV< =WGjUI0ޯb];dhV\߬_;xRjsY,`4flt7[&@dP>Y ~܀p|Y]Z9W\on-̓55-,P-x@y.6ת FJ)vs!n0d~,1 ( 4G`e7<lH"cAɘbZɃh/.wN H6lO/*{j8/{kHu{6{3|;$Wt|=pg||Ժ ?if㯞\2tR`kՏ?j teȓ, uO}yF@`W Yd!#Fy: ?O6KjwiPl^dg;9"'@e)׻$i6)B:v^Wm[nI[S&@:nixlg;MFFTQrx%LQpvX t'vc̡݉> XqxXjMôqO)Kx(kZ!՛uDzQV3X{xӜ(g@g_hYl̓H ad51򚤉?>'Zfk+e:i o;ù2^IC=;۹Ҏ,֤.FwިʕYb5+n =2?A,ԯ>va:/h _ :~^yd{(U+ЊSGnsVd<5xotz97Wʀ0?a!3Uȿ kk$xh4;Ǖh uvSg0oŽfƨ =و}hh CcʁJ5Z4 F_=+ɯ16#cAaGx5bCHuW(`oL`Rzfcmdr5JM`8DؖHǢ^@8wF4tm9zCbL0R)ڔ-9@)OyG/!N`r&8 zD|a-uvGhJ# M9Khrc?J ‚>U}XWY@6Ѣ^C%Y˼|ZO|oްGPaN$zyWRL~y0T;C D~Q2fw_j]H}Չ"8OT$OlnyTʋHƼaW ۤqDԖr5 T/5}rU~(<D-Shgn#S)<m&&=D~=$.'CKvhMf߽:l+lb\S%׎S~|%H? v8Mqv3OEɛJ^frlI+ېmOu0oMΆNa? ݤB~" ODQs?3lԢ{t~~ZZ?&s۸j22rje {_ӵ5#hJm7=ק'I{Qu_.9:oa(+9N_m GO%:*2 I?_d6|U7XKX/WNmr5Qp%Vqw4)K4*a^njEfckRL‘oR°J!( AK}慘r,I`!rmR+|8s0e뫜cE]l!&] L^(E zf|Sy(N%, r VHtB߆R]% ! rGg<N7͠Qz@neEGoٻWǧNF!ߝ%rx)F yZ<]Nߜx\O3;Zy\F01-xrJ9UFZя%X XuCtImhEx-jTJ@3={п\~6hyN'3ɤ=ͤ O1#"KeQрzYuuMݘKN~Q9JKzUeZNt0[ahacO}3mj&537S3~j43]+wX1V9qoIOF|fV\7ïOyD#ZcU&vޖ=|h]vQ~4$m{0 ϯ7K 7=T\'1v}C:VFV4HW38I-j=V±pH 8 i~-c'4W{GrR>?).dvXھ`YvKx'#y4+Z[ui.h߄ u%3T /ےfe8q, a8jmCen x<5LSg=2vL_Uү.N}IKFe)'f[чצhW^Kݿ߶{C1~s7۞z\ԟh8[M==\~X^TEȹՒ[  /Fv./iDI.=t.xX9~y(;Ft}Yv{A2.;kG٪6RW͸4n?|},Ȫ[dXeTSP!î( ARp|HĕԵF7/9Ose8L7&Xpՠ]*fT{>퐺L#2 {0wGUJ s-욖mb<&WώUp <a~ZtrJR{j#ICs6 R1{>3>'1o YN541_:}FqtYtc_eT[zoBU ";(^ tGV:*sr4y缛M>wqsj\ڃHa^[S}_hj(@N-BFU'҆bԇ(<7VL]R!sL/'2e}ּZA77vekMIh^NZWz(e~+wwm !?L,BXyRK}c>}$%~el::Oo.P'򄢒h'0@Hg蛲Pػ@87"EHt|`LWm+drvdlU.c? rd~ D2;pGC˴m;0 IM_PpLWdsDǗ2R2{~=AB1 O-*-Ph#µ|a Nx軂:,-X)BL~=۰DGbLrZ Rre6 cjGO럎{a-xn>{Wp/9r1= fO fVZ47A\UZ?){iQoԐUqsfI9Gq|6a!Jӹ ?U:˕{aX`"%}! `RaH'\!*d}+CV\]YTFc*QbMGݨx &SUdR=@ |d[v`}v=%Tx!E˧!&T\*SZazT8L(mb..R]a@8 [1V bXƾ5F}dYGJK=d P-@'ҕ.lspf"T JWZ6Ҟ Pb>2(E-I;#pmo_[2<U\+SD頊1}' $ġVawBI͔&5'fO~m3ֳwDKʤK< ;\0s]r-aI;0D mj!ױ)=;b5>9p?,*>åd%ԉ Sϟc2lyNGbرԘ-5_*`2Um7dؘ.IU6ET2QEOzl5Uh 3ԆW:Ff$v8yRjz2`| *2->l/w'}\*HN!StsZ~/?d!11ˉF&Ct{eRTuMI i g#5smAضmʤB У4$ Iplq >qx\ϲf~Aq[8fćPc>9Č/Ԑ :h`7>Ԙn>s,n`jE;nN% y `$KZόAC5ݍ^-lށ5zo>0'u%B4ʵ=9~^k sC \keގ%G$Yk28I8lP`gTs2wQcN@((PH‘0 <ߤ4^,WxaIJx6 T ȓ~q>x+yr ҉Ji&9{Q7c#"+~%E-Ifkqi770kSö94SPjDP '^t]Ꝟ3T wԋ.3 HO$VAy;8]wNkYt%[I4d rH@ S|Iv 8"2\`K Hx{![f] R-7hNj]D[T۔4//z./W0 b+I 2|Pbr7Z^GyU@l;#6`QM٦X[&^>u6:mVոMH\MSH_B:9W']] )a`ZgP~E5+xAmjf\{1y^hVg|GS&LENvޣ+ Z %Ug6yfb r9]gYX2h9 vA\/swЮps-I}J2fd'0,s~uG$r:ot7&.:+;.p8EN2ĤM/ÙhH>~(< AȭJcݳ ٳe]&c!2|GU}RgZ{_nh69`q$@Ef~ (%%{%-K6d2|U)xeh+H~oaDaE}cB``Es8]fe~3I`_4" !BI/$C1ʋE=t$ íuR:Mui?MO.BÌn4"vCAd#ݬǽx-|NUkgXĸw(dDjqj 3= ?B`6/!?FJnoᏭ~؜YnP(O`*3ExV,]!U8`| zawZ9nY'tġua7W"ޥ۽vZ ?{eQnܜ= L( }8p~|ۑ28|w%ɷrY(-3 g%t\]>`E8P7L[˛iǒj0;jԃLj߰y7fuȞm:.tU~Zxtc.1YIVh#4Ft1pN Ue>86KLI8pO#4*8p*ĚouV@չ n["V2)y/*vJ c:諙oxP 3CQx0q+oOf]NA'lD˝r?iނ-bGn(C+ac8j#:C S(:dL$/#d(E S8#J[6^K tMNoJ5e\;¬$S%$5*LFΦn4J^bZ5ə/_A(ha5GZu&{雷 nFFtf0@FaTbV.?]/".>['f~)ɿ< W? Yi=a[[#W ExBe+# ɂà Y^]`ŋ%z•@ ) >*xL\݄ջ˖հ|[lxT+ĨRF~}(~WQYdVx YEF?8*OnsU!:[RlhoFG_;@m?OzQ >[|L{#LQJ5bJr$(l*o a,{#/Чu!~ȕ+D<.l~s:͇:nill*٧V" 5 K}p!f@)dNb*Pq^𷂀}X# #Xڣx^+:SEZԲu 6gx}'_=&χ3s1ClQD+$̧2=ZdyJ6ޤ&.Φ&!Z37!cG3{T.TּG$B `6o俛|tg<SrQ_.&"[VQ鐬]PoЊN >'>WoO;1 s%m>-Gc^nFh\kl"O q> .@; - Yp/Pb^\ŒiuPwCQ?x30\-F7 ,zh $\tw $I(r[1$~&^_Fd͛%BCR▿! i־dՋevSٖV,p<)˿{pkOuX#1(Ҕ(d7'Sڞ=wGOO=:7wS7]z=-S5UʬӁι.|}cƜQlGϕx;[lZ Y-R2j 7d]Ar;Gy09痓eRS!#6{ms$GDJ4JxSsh&x6^:9_Ĩ"_7sngr㡻:>DL׎H VgTY_ʫF,ㆂib`G,$)8YE+^ W%>(,"b蕔amqjI&4JjK6).S8˩%}P9D7Pc6o6_xssʘŮ 56<&t8q`rccP(vPJ|`oy#cSR/A##lX$SUSTF,h@UsWG꣐TG\\L'T:.MQgI?PueW Ϻ{2ݨPXf{ J#(]?Fep*]L'zhBI%k´{5 ءF!W8ewϪWidL fYw.׼#gife}F| x#9W#JyW*f1IL\e(F M e,zdzz r 矿>{ˆsFa$cgS,uSIHN#N 갰 7P'7~=x..o]c[(_mӰEv,DŽf|M\1^ $wc.QQBI^"S8I ͛"Ve%DOpvv'ݿ(>Ft۝c ~}>O@1rw0 JK~pQ}Qywz?1xLII &B sC.=#?~ֲF ZՅ[X- q2+vfsVm\|[}xP|xԖ#Tks /y5,ޭ䁮%TF~}(~WQYHz귣:[EF?XQ7FPvk_fqe~k¦]S%nH`ԶThfmt}$TH*!cI iA&RQL\X^2gæ 8qJQ!06 VD+v#7<ZÌT:A\c92:tYGGbVf:IIzkLŠ襶>C8ή,jĩLji%5*q"љ)EbM0o .,vY& hjQ9Z91P:Kd2AKhߌ2Y‡bp%b?M< p atR,' Fd2d %a*E;+g8O8c0o1]i<*ǟ4wa9玂I J)\%9(PG)o a,C;>wȱ[VӘ.z5x]%,!liy44/QoL]/{cۅ$뻣BxpwkXǒVsw`FB4ꢵܵj(AۀOB)S&uY)d$Yav%dM r% 2z{ū0F Ug*0μSwO6rpz!ػS@~rj{p{ˏ#3 Gg磰0:BX7ȑa{w|k>Ace|?].1jR00A.Cksr.4tY(7')7Gϴv l{m\ |f>v,Cu%nQլ>pQ.!"Q0`>`nyQBw3 V# zc(Ľiv,֫LF=Ȥ 'pkLWW/Q˝q:K _cqc-b6@^|4,s{Z/,͡AUTѩNv:doZUel7j +rT9y/3UMw9-%t'柼tr-,TP,(Ahpz> Bs'Ah৞ Rʘ2f?*yp뀻*_^nbtG ”"&8q0w^mK)J1)K,76 v}kҀ=~>]˵RhMW8N0+ zI4I &RSQ嵳)[-ٻq%W~Z ؙ(x۱=9%bESaUjj%t\\^?Fܘzʠ"5ʪWJM wW&Ǥ*a f0V-*U%*P֒YBSl2j9@jc`F+8i+8k\k.$ՁUɩZ} F[YK#76h O7o!ޢE65>>\%lDK`>Hc=[vMƈ=rB\×o9 \ĠY Yhg߄f~Y[VtE1zޏlծ;Bv-?sG Z#DIsG9(! ?9]5,f'FpէFh$F(EW]ўz멿_ GC%O`!|[a.'s_YI;&0RYL vcm>_)|.ҚἬt!y~z^PF ^䗨 jNt֓&亹jKԔsګizD֪I%W2kZSGhZp% bb#ZnF%?w/f):DWDu-NW۞.悙;•+th>wtu9t%4XֲCtm JBt%&=]]$]Im+!gjҙ!s+DiHOWHWpҩP;ݱ]+|عUt]]"]ixA, ]!\ƻBWV} Q*ҕ1ରV3tpug2 Z{t( K+kbCt;DWWw&vhهCXͩg@fOjoWzZj(v e)vtةgD0M;DW-mm=u}fh9w.B.;DWؚTwBPjҕZ!I;CWWD;]!JFz@sj:DWYĚ5+th!NWtutDtup ]Zu,m=]]]iF_=eHt ]!\ s]!sK)[ѕ)-491|KW2YHMdLT~Ts*kNv:sh)Zfy9wQMG*E~`!sb/UyKUKU*SRs]+A UiU]R4-Uu*NZrU+kɭZ&jê!Ri+YoC:} "MDLbI ޑhH9Ⱥa< gX +L6w9L)(Rheb%6jr<-qpG|(}iS~kf#6" H+=,[zuO襮O"_RVc'ҊmQ_~q ̆`ayAԕGF{rq!V;J|w^t: ҖܮŸ[<>HI?:a=حV4mVy=k7-,^nٮ߻eUg<;6N-1|e7gl;se'wiX3'|CһBksfLǭ h;,l[l]ÁƸsc(ki)uK󻹺,x8sa%qf nኚ+ܕ@o?(>./;4mo'A(<7ԛHMn+NrTs_SZX&_owXz.ZuR-A%]*ކDwRkAe~W2,KQ*uQ&UL;!EzAt\E:$܏t J#p!K˼||om*%SYrcE,L|uM%Zm*ﭏkG|?V9ʹڍ;,NdyU1Q>WLحvͪ =(X7Z3-Ǵ\sUل'j| I˹c%g[sluU4Nkt? ;Xg{\O(mqfxͩWSN_#HD3ĉ'4B-so@W^qy6ХqࠣI>ld)Ό@&1-HjVzl#s0wNx7-Η?$u0W<'l`^|ٸ<;l07^g0G.xM/?/[wY#OIq䟏GKk}'ۿ绒?Zc_t%Eyh>:\9,W{)XBYVTO9V`W.jUm+?ޕG{77cOaP_緎Iuct1*1CLj1"'2 )HVgg,u;+SS/D,nZsYʅnoټXeJ̸4on?:Povaeo>uu`lqb\ &c?]bn6߫&"/Ov!"WXZ出^Ly "]^ Yh.Y,fr+&gԫG Nu>~l9tfW GyiBZ&Kms}ΗQK-}C:rX鷖/c.<-xT5pካl>|S?vh9<"onMyT$)_\wjڳڲ?2%rӺ *sEk}Ֆ1};˾:uިi^_>siZ+rBw ؍ x0 Y 7,ލF p;$xZM%f $ .5ezc`<`r9v3,t4 btg1+1qoYGR̮w>_eus+/SD,3B &q7U+RQd:Er=^ b9sKyPJpfD85i.A[.(93$d迥cY[֒TfIURci/J6|''t y:Opk7y۷Y&&] w+mijE%\SύKYakAcZc 6_S}X١k&[ LăJ-[̍b2+V*?Oi ơ`>1%_Íy=wg5l8%΍ٮ[3vWۮMF4AA1{U\TNtF+"6G'{dnnuNَ !=<8Tu eS?'V<̑h.?ImtLx9&%8IKJP:W2SQ.Kk \apMpԾEMKmO$8o+)9 4Sh2pI΂6\}JJˈAZ#+MQ7!Ֆ.DNM5R8q$ڥi`>pb"$(珽 5M%LHJ}jP-2|BnjPbaΙ6h ʢ[I\FRS*E E+DƋhA[ESE1Tt)߇q3I(%Z)XR2jaeEB2kaMjqPwb ]:!( 1cLQf13Գ, )Xt /u(/p@)(5z-0=CABs2Bh+ i$q&@;siS C`Z JrL;YIO`11C0P ,"/p@IK8K\LG!yaWTY<ʤ^EjÀOBϛ{!P\UjL(w1RR9a+mqkH\fihk_2jM|*\Kɴm%*jh DHB})@C)6)22 F@$`a=QEys#HP2dDEe(`=%rN-"ʅ` . V%K l! vG"̳N M[D,r,U# ˄.% !.'DlTzCes CI8 d4 a"1d5WXli_0Lֺ p e4DѕL tI,VQn~$+9d2و4TQ"]EVU0y;h)R6 i+K z*RÅ+iRB@UQ@ISr,/d\[~,!12H` ,BDA %xMa=`A@%o7-0wfS鈬qk;Z(236=qQ/{ޏb^ueŬIEEicv0'Ɛ'8؟S`kL|6/pF?7=*NV7Jtϻ :cMg, `!m2g(LtnXXi0Gu@YH^' _2/ <o4 萰LU$ 8yG˳T%7Pیl2(($-t'AӇ)׳>QYɘ[ b`X_Yb=#yhGd߁ xQ5A5i]%ྌ H :!D(/-X(mt{d@;۶`@ȃ4Dg :lTѤ;$,:U3*0cAJLd),tZʼnE4T-k há:UL򒵱- AGA~EF##Z.'۪̃0ַHԁe= Z$c}:]2*5aYY!:\0o F)8k@f}ެSHmJeJcl6,ˆnPX@K'01'@|%(!kƐJ9 : k0Bg<-/!t\Nuuiyd&^ P``;@7).ffݷ"fg6* &=;{(xocPuVyk1E3`4vƘ ʯ3 YU:]b wQ"DNwB%BB4G>=  Is$ڈF(}V0wzD R <(8 \)wУ= BW7dXoWHO;ȊD4)3\D9 \փ}[_uך-O`> KZc8)dR!*Ɉ3u5Pp iIb]19Q aVv rQe֦.GJ a^rI32ѪUQhFw-7lxd`uנl:n/Kb lL.wA 0 gibm(j`)[pR$H+=/Bm)<%Cq{֨b_XUCUruEDtCkfФ!F\$?캐bA@$DiΫ\|ղh;6iV00@5nGA̻뇳vaXd>73 )Uߺ3FA !6{-"_v҃oLWQD>8ĕj4$pHkٓn"C$wDG4e;l||]PGj]j.ӕW YD\k]Lõ@>Gx(TD +F9PHK ֱ,K2et1%XZy;kD&@s+SLLn:.aD}Lߝ>Cxw'=P xm]x]TwQ.*EŻx]TwQ.*EŻx]TwQ.*EŻx]TwQ.*EŻx]TwQ.*EŻx]TwQ-ޅ||mz4kHkٓ!cD~ku1fn1_kƵ bSio4HT&6?!ߟ޿s]~9Q-?]'htNKP/v`GqжC|[}lO7֠=G_((hU=yqگͱx9ױk4R?0m?0?0vŚ{xE>Wg9A"c*g ==*lj=Nqb{'8lj=Nqb{'8lj=Nqb{'8lj=Nqb{'8lj=Nqb{'8lj=Nqb{gw"{[68ul4qx{qWv=Հ[ןhW%*P,P7oF?9B.Pȟ-OM@].Jrw<dy|Ŕv3&>ΚeU]/eaĿc^Lca -n?<0Y?6GU䷣H|o^3pe}̦a?}slyVڦW| ;t,?}/`| YM fAl'5siÙ^ŏOk2Xy+?ðDYxg܍Tz|d{, s'& 76,2W6!tq`PFaҧ:*ֆ𴀸BV5|\E?_4R9y%ptv}Hai{iKZZJTl]]Ӕ2˴):00Kh$Z~:./xkNYL?Ml\_xb|ڄIŁZ_zxa&=xO,D+5%<? qԅJ[Qz)Z&vˋÁgu-T(jI;3J 'F߁]'>r>i9Em AHe.KRx*2ޖ&|0+p:MUv$#W3נWaPbf-؟a Q=،?bmW#0eWv6R & 'V(I8^cǍʼ%> _ț 4wϋOӹDZrevt dhw:'+Qԯ?<=}[\WӇ9g{u7t5;}~52ߞZdH(qHz_+a-ʅY96sc$$&~5 h]@@;B7Vocn vN.]K1<ϵ ]\M/ngˋ;u=cORm{02;UՈNa 61jO5y}P||M)um#N7Χ)w)r) ?NJ^EkG/zbW;Z~Asl1`K@h?\k~*O g}wotA~}26Nzt9hS^FA2I\`yQ7Do-MjO7wDÀϗSv>Q#1ϙo#t rEU6yׇt?ǃaퟹ6]|OfuLP۾J|'Է^ntq`;DzWեیtGgCd]ϴ{ʻޕE՗aÏ]~[5-e.hJ/._[M,]Ue2%u\,Jehl {(YX1X$1)Ɣr5FX #>Nmf} n5ll'GNii}Z]ԓY(VWVgl%-~+و jC`. >EZ%ϽvR+* jKΥ3#rWXq3wUj,H+"])lG bAWE\⮊UXrW]in#rWE`F㮊֎]inV*R2C+#ZƴvtqWE\3";{t)/]werL [3++$I!8ˏ+H&wݕF1+j4`֜}"~w9C+k,=宎Xzktu:JZ]'8 w]x!RNgeEOd)\?>juҬc!<V+YT[SG3rUA.Jt/#/Oq{N~?,{2h^' O8\dy5ζm#_7I},rqpԜ9A=RmJYp76ݷ‰-Tp1NA0c4)6<)6SlTY~ݝ)6 @0X 2 Pl}DUj㪃RN 6*\\Fd(BdBmEp& \`#ʵ6\ZNuq*K4D+)%!LQ-WmUq&$\`0`pr WVv\J.z\uW`ZS@vK ZP;cWĕQN<P&Ԏji;PcW]ĕUZ&(X]\I Zz U~#׃+UՈhRj mJ<ԲNRILUWվMO)æg oLg7o銳}˳_۸*7*P"/X]_d1Eh7׭}јСEYku(Z&ڎYT)Dbie WLAc 2BUW!MbA&\\C5MmWĕߊp?\wjmUq% * \`c‰]Ic 1 R dL eT]ಭW(WPpje+PYj"2tlpdvy9kԀ >Z0?g+ƵX·Sp+_`D"#MV.IQ¯ޫ r73*+eYY~Vzv5I|?`էY>Y .*ӧHC\~ {z)(BLn 2'`߼s;׫򻜐2 M!c h2!ۺk2/ ޿{=< %t$`IߤkC?_fj4OZNzκ:'=]Z) *%tzje'R.z&<Riwz輪O$RR+8Ry=Ǭ'T> A#=!)%7't\G@ofOwCU+ guqTuet ֖_lkզV|-lc ]HS"*t<ou/.<np)nYǑK%K<1tR}zRx{ܙ>~{_U Δv=o%M&mW{nAr=Zɖ29rbkc*NTSw"MؠhcyCS#T6l%n,VMK _-d4! ;q4dJ7`Uo#y{ Zɹ_w{o&|UO]~zD% Mjچ}Zj)k z*nkjnzt+t>Xi>o#Ӊ![`֪\m V:y VFj*-- 7UtEwt]ބYo'O,օ_`t|;X:pq~eD_)xgz&h劤ٶl5CZַ*6\iG!,D5rWVKP"i LQngR:JO8fDkK6J+m(BWe=W Xp++i(BʴWWĕyd kf5҄+Pkl;DWĕ%r2W 4.\oWĕb@ *L Ԗ7;W1"U Lϲ&(ؘ`p$+T+ZjG]uWٗmz(@0 `pj5k;Pe鸵W@P,\\CB+TLWbӛRZq\ՓkRhC+T-+jߦhb6+-2\Z[+Tiy 9 W(T׵ BB}eq*ۖ;Ջ pyG(֓kWmz\ a  wrU0B dE;W XZ @6\Zhq*-C]ĕPgerFCJ{W]ĕVJ+l$ W(W3D+]g!yW(`pruS Uwp)xHD?pbW(WPpj0q*qVlz;"V㪖`K>M\dZ3Tmd5pe{\VI&CR\V'UZ+IL+[1uE^0SzʲO]N~ZlKFnc~:ͩ %[T-b̲{߻]VʨR,MLT1;$>5) " 1B'ЙD]uj·oض%=^ \974Ms?4Fn/d;H};VT[ћqلLrނ/j4ZB-6=2l)U߽/nq巟YveQ0+} č:"46׶r:].uojx3>T)#VE^fP dmUo 2 JCM/$Qm6Y"MM3iJe$ X"M2BdmE/ ە2Nn0-}Y2ňD|ߗ0SKi2FR3T౱q&\zz&Lũ4KŔJ۰4 l~((>u%۸me. t~kY0|nwysW)LnXՓ||=\]e7~vudD0B;RުQ:^\[$~zӗvr %?/XoK3-Kpq`y[}%!+i)PS=?%lވ&wbc6*MVi N~"aZ$|S'2cLєjAGKZ(dl17(0|VO>qj1LzV14u1`9EmLa4JIX0 c#м߈n_jhґMwvWmpFʇgyuoݟb.T /Ҧ:Ɠ0>|plɯKdn2=/awX]_懩΋?A'y9 =y볜D| ^bdYTE0Fs2'#\xe\/G~bg" ,A%ybU,uBjmn_BG,5TG#?]9Շm8sOP EN~_sd]KPgW,:==Y]OoEi0d5Y*_n~`rD)uy++.`7TJMeG|wXcKy<&OEr{S@+ 7VO!I$;]>tmަl*o~:5#l/Iap`d/$Ln1֍,9j3C[o[,ٍblVUt3њqʧRN=qT"aJH~t-@{htsNH'++k*0'1Lǿ?~̟_?~Xӻa (zX@MmPC@&Т==-"N|9i"Jmu"  w-1pO@a%F)MW)/R胴9arrRGƢqQ;T z*V*F8y>dSL=]tH/4qt{ $;xLp]וKe ]ϋ⣼f3e 5p7*@R<"y&B6&S[˓Vc ƭZeXBhtT!h gU kQf˪#ܨY,SQi0IcRaSh2IictJK"ְ嬱l)gnwsqY3Q8h&uQt M8󄭀+)+ X`dH si)+q"Ф7;BO?w)@+K)8 D(C#H^otj-6)%f51AphVcPˏz-OjyҊzyp=%8)^c׆'` uk5e9s!, oI4qT^ލ͉]64Ia8VGJ6*l2{Ȳ 'J[sh[#1)|M lX afKa"p*쁺0vH0, @ѱcƐt46C*aPz[޹+",&E^BT $c!u],֐/cYdC7 wbl3JH_=%eA{)DDeZ1wAl"&$S1AbX'-uG!ԁ%BB*3מ՗U\8n{t/tVxPUwFP.U_l 1gFZ~ֽGs_r۫vKR׷2fbRIfް@9ͯ4~o58b-aObj q 3mvA[ӂ6F|2$^04Wtfr/LNY|3oUtEYA-+׷C;+鑘#Wc|Yhc"Q'o3Hh!KaP'Ž쮼]P3I>H/]8x]6w=6ѽȡjM B&ByN(@=9#7XsK[wdʸY}>]7g8*Cjpsj__S6t xTW4ɚK_!J!$t Z@N)@I$nȜobU|)_+][J׬Eo۬j\7˄wzE Ũ`%-sѬT^r'yPcm&?%J7V(oUIXRY @),p,*4R%1C Id7 {`Dje$cFP$aAq !2JX\1yބ՘e[Ik%}JIKb&REm@tFݣN86[`cM\,sA/N+'F|iv02|Mxy6HxBcp%*NOi= ^Éutm\,% PRj, [C"s&AS"'47Hъ:Y6dpƽDkg4هg}eS|M3}tX:'öӹ "j$Fكg嘝W^EwoXL_#qԸFݮVNjV~zxM$p5OEwЫIH$hR`Aic5YGwoF@^_\5!F"5CbuNziYɰyrkdBʘNB81x O'O9TIYj߮ouo;W;nkG~6~TWf9@ Сvt՘>n~Oϰ$\vA\ˡ/} '=fquHsǎ &!*@n?ָf]c8mW_輾Aa,5i#@7hd>w` rZZn Yjw2qہ - Jv}3cw  ݮ.ҷ'7@ fE(ԂjP* d] "cɀFFXRdDLPuX`х,H:Wi Z\\y^xsYhJ`M}ӻ, U/2,zͩ 9趧 "))9ӞSE11]6 T&Bh5D,=&lE·Gn'37ݤǶQ4Weގed 䴳FFC4:᭶&H\>]`P*d-yHJeJv ٽ&U/FzڬI5oc7|u.7h)==$S/ 14k^K^!BKBH݁U)\FZkY Ɲ˯w6;A:.˰ŖRa,9Êaw;f;g0\r3}HRedzMA2uA1C'o ZТ`'4*9UCGtL(6ǔ6^xA2II .0 &:G0s&imA1{?ssm/uj*\< t]ضbk ^$, pSY)bPVż ذ(WDj,fGaU^2't2._ߔ9cjR)y/7W1XI_g?^[NT͉Ї z %Xt9OXX EN8"*Fhk?qSOl٩'`4Y˴&c6DL,I%ygLXtQ&N|Iu%#VSZЩI0 9yr.o[X\%K?IMKn5yu-+rSIǼaGVu&c05M5Xqcu*D*f?&k]Mz_T$S+ɤq-삑VsY2U!e >847e2)k.Yr 6ftXɅ^Eռ7B e]8Ru upMg!D!e!x0k5f,`k1hFs+%3y4gX\Ətc^}:$4 b" ц{RӨF $%J#[WA$%bB 0' s~]`u$S\[H_}v6"j+ݽί|iQB[LqRvꩢ7ۡyTB_jB_B})n@))& I)47 ,3A\V񹘿,]_QcV~ar_Knz mOߍݤ?n_iOxu@ @6 j-$* %^FHHLBDN.t˛@H"88fJb A|@N@`kC:„{iLv/e,310BB23uS/IN Y$53Gc]XKX6KZ,i=ŖvZ}ډ胺Gd=Y>hp ,>I3)jJku(߹v~;)J1u>B`["Q LsW pCcI=ZJEV7{FOZ鄒10\0PҠmPV# $0Yfv#]MgfC K/֜$p# fV%VWp#- Qi06ypf#82:Kf^Fv&jg#YCqʌH!*HY0cíkÍAX FʙAPI(a'i"$"M`NR#*#+TT!H% M;Φ3zi . JHB-1Ry)W!pYiq1"& $aYǴ#wOM ~ {̥&$E\ b5 7XlX#5 gr~BrYU "G0A)gJ((P2NHa;8h("dlO<Bvw1Ӕa9%B2jfB Ȇ29#@Ӂtc_p rR>_¼ vї E0P"hl'3.|@YQ,;iừ6OͿOT澮mw mVћ+vp5=6@SAgvR92;ΙR~VInJd,S5!Þl= 98?MELƟhX% ,<1waHǔ{gqoR0#4'(xAcG% YJm-͑/?tISVmV]]䰝ZAS2(]JV.;~:-)(!林/xhƢϥT|^4i&8=iƥOůeL6[-]wG0Ʃ3P.o( 5>i_ku Jh]BSrzsNgmgn\ͪChYBˆfn7-~獷wTQ?Lvy07\8y=x֝o؛p=,A#/]omfTkrANDG'&̎zG\",E׋)M޸$ _|ZV1U^Uy5eIIM޼ >ygoc|#(ݣ&L'ωv7]?fnh,suJ7<|^#%])Gqc%VRW bx+"R V//x˒ ::ߌzՊuǰp4i!Iuh5Z`Oetۨ\ep'iuBVeڱ6p8mfNvzbj1rjCa?OӘ0*0+xdʁW }Ɉ!Q:)Lb=G\x ΁cB)7j<[֑aYqGkZ H;]6f){h9xyqg'c k2l*7ڀ$~h(꺕飺Ge,:uy?'ʭe,!]Znu OvAe(ogNW Lێr B($<3p|HbaJkm$G]`RKj$ֈΐƜ=4 g1 n\5` 8Xy\`X6¨3帓! *$"L9*|%E]rBesФFzg "mqAT"ZNP [^AM~uǶZֶ*:6kv\{p?XR,`௦l0M9)<Hn!m?6ǏCNF`4}Kw8:~eO9ڎU`) `j5&j֓F97wm+Z~Ay;`ng-ϣF?u|TG0Cw`fLG  K9RKl:P0Yc@JJ0 X30ե$g,uɿE+p`) [zV(v܏(Z҂x@gjB' 2!1`:@yXIvFACKӝV=19A8Fqj;cW[ 3y3}),z͸<ޞ*me^jr\Zol?om=V\xWxdc.qsbϣEJiA{nR8t([q/?/ˊ\=i#@ $~硽rz'r-ɴ%J힫NzRklb0og[uzaQ|n ͅE]r,;i)=$IKjn<2g2j%^"~P"/'/^ ( NynI3nx674iMniRrԯ@|Bn .>z$CrWQ~]%q`U]%)iqW记T(0j0*zB]%).!$qWI\9te*I`qWw讨R(bb0*{ӤUJ]1()4Gfdӣ'$)ߗo__ĩ:" ^1MRgLE݁\AF\0 p,q_wi_G!N~^K!LM1w Ca{iESCbCqGt8. TVc@RzLitP!ccqeS$k0~"C 6IWRk  :UeIeSȹ:-2NRG!<^T5c@"Nf[K) ra'&50gQ v8)g9Xх,-DRF=4Xg0BLj*v4}Ki2}W!i։zXJj4;m=nl;B,a´%J43)M Nie0"6`cD\f/LVӒf&Ͷ٭: ,j ~?39nI\Fp bؠ" Mj{[@4;l1wVAcPiiG]Y$L8)sƦPjq.zͶ ^5Zkgp.=>kCWni-7aвk P J#âQz!rl*TȀP'"(@8k6JE XtT,nvTx,` Hy ]82mBztNHpHъ. W? .=5dԆ*6M= dvh^L G\WtTfՒ~Պ'ӷ5+)=O%^і\I:03vYb?6nf1MrtFhU[gs=ܮ_v\sp禞c{a6^aOC:}a0.\`(z7{7f}?1$#r+yq` ěqW"~b&o kKR@Q@b!HfRVnu K~Xѐ kQ݄h-l^.[WݦǰS[x } o/qU#M*Pq4ҞP-bVKiSFB`,':Z[P簥FgV,8Jvh >?{gA`o^˼&w?p#TΘ[a"QL JhD`È/QLܙ8sn1ǵq㟳;h9~`fl)9ad퓱y˷)@2Ci)cdhB 3 ̰e_U乥!*9LIo N3GVPh!1'3$΂^|v5zɝشI.5֟nXsy V*ۖr귚ur#VĚB htM 5NIڮ9.4NiTv8nw?&C#"ꔀzM!X _^1Wt1CިCJN":nPVRN`t[7Hzg "mKWKZN|)4HR7\-k[5_;_}^1aŇ؇TUihwyd5ef[r`GkuC6?rSEgvg S22`pȸ r0 (;BњH8i/krJG6H(y49A> Fj%u/f3鵞ƕq6uO8v?^PFAXQ`_3d077g j E@b|}f}|6rJm6@ PPQ`*Lu)k"fK]ni>\dF 0, 0ҳ/P{.km@;{҂xʅx?CV}q289AxW_m`7MԢǮ^9%57l5 iMWv/?x듗/֏Nm̯ _ T X-kx\],qouŔ42_IŮDZy$]}u'/TZ[9IA_碯e^205/'o {w`w'e fQà -SEExjR^-H fkȴsInܴ:s\b 1oߧ ~oMw{MiWэ{ؔy0q'yeHoɆwj3¦^wߊRrnuɄo΄svEŽ: "9y-I;= 5tt *%\ ѾTk0/%&&d\!%:?^p5k1u G`3bE$QL)x(etY2=չNH(m%h0B0li&`eb+G^(Ժ4@2s"pXu7\z|b(8CNGTK zV4T< 4 0f- !-7g( bB#m Ĩ f{u+3}cqNK>-T4|&ʻtmۅ Lڠޟ9˂& ~jB t9ׂ("5(=T8Ң`P68 (;y9>}5@,= jF܋ u9&B.uz_Bf SF9E&Ę(0a0 J4 ^]l^xgPq``L7eR9H]> 4.a1B 2'4DgŢKLc1tQbh!y$IFL\e[ŖذߟHm:}zƇ.]g]gs3P8q!)mb{6Q 12`A4PCǬĚ"e-߿CEʔxHD.ogg;;EƥVH-4 ?sd N?Ÿdɻ,uh(#K^Xm9xց qzbR0?ݖ"=Nʒ|2&䟒$&,Rde RHA%8s$kCJ嵮#! vT&3'̣ MDijd1pM *2R@\H-3]Z[+~Bw0͟Zëpf0 s%.~hʊc3$YR}TǷSaѩ 3=yr}eҤākY%W'Ɋ=*[Qp-*E0&NRrh'q0<9mV3@#)99X\ 1tAimBG89oiWP5%EuQ2aZ7Ϟ=^)Ssg`EIKx7D~}b'vIW KS0o+3&ݫO6uMZ^5 ߟ^,g+.|?;:`VK.l ~>7AC/`>cFҰv$PG:Y7X; pvERyMlp0lda`h1CrIo|@ vIu\T磎JL8 |6<#^*O%)ּT3qK5*NO_W߿|}Hዿ8|{:& 1J@5mp[pm迿ymkho6`ڜLS^3 Qn^kֆ|S|A7r: ]L{*z4 Ml~xUoT+ })IjD% nf}yM͏>#=TeS0 XYEg']Gẃo2۹ -L,ETbH.']_iЏOJ3ԙB@%d ђǗEc'ӥx42M 0r9e_qU0:ߑ1(}hF}q  >n]U{].>ubo̹Uul[=˭2^k*ƛngli80[w5g8 'D G8?&'ڸ5(2t|uQ >(Lc1>3]!B|}ۛ>FȃZ#cM7&g:89j"؆@Wm:ch)3Y&,B8Y*:#ř{]#hV2,O/of,--?,Q2Rr&s9x'=[;"{]}cPi LL;d 6ha3fCWI`m&(8fA3U,I@ !E$$}V|JEC!9uP)ll905p>ϢxTF.[w%X5: NEsdmY|7%71H+_ڬ͛Ri9w޾/JN-59ʓ|)K.H*`*2qAl &͵+GlC-[d5F8#ҡC7)-ɀg a1}L (rAj5 L&۞Y8s̢mȀ%8,+m| Vq )2£DC̔E-PبIbI|HY%&G}vmC㱎Ǿ"s3 $wY. #Ԩhc$sVp"Dhl1wqZii{akaEOQݝߠЅRzwc!vfXbT8&T)l ukMU *EZv\ޝ`yL2&@օ,1י됴sV)ldcm`WtڢsB87AAd NN8 ݥJӤ_zyrucld{SqUf([R ,1D{Ck$QCoV9J:qnk[/Wj8㳧JIzo\pRsʀ|uU&)w\Gjs*y4B 'BF^ m cR(ljP2d|8 =PyWvf, 76$%oRyEOX0^sb63#:h-*7]L+/EO\'R2NI[sޒ98K՚O$aM-, N!& eXYKR!G\"ŠxK;5JR7-MԴT>%BaF1Yg`-K 0d˖1$,8ȪvD4dYNAܕt7{D&T}⚩`R 9Zld M9Ox{scl.D!w2wǸ z>*5hR7]YPWN1$J, Y RjRI_n_Rhﻐ KSm /ވ+W/IqqDUlA)uWråiխ  \Ê[Azqu+(=털WWw^z%{5^ ǹ$1mj&oǴpj2F}ɔ4>&2MPMoVA(Gkr2IdfdF:<흞ӫ3+m!ߊda~(n[ i?o5/-[LSΦ2˒[3ϭ#9d!YAT; jEnX;(*PRh! ];"[խ6ԙ\gpj2ejϥQ2-\@IR=r ` bsb]w J@x\l$`foU}gm/ u+ɝRb-qEE\OW%~bW4sUXW z[ Tz3ٰLݥ.KYR,^X*TSzΧ?IQe/=O!,@{霠_,^F$Lj9Rٹ7ܩ'?=5wsa{.p'yg8;CjeDRG܂W$:C\;oEAtH )H ?=]&oѲ]ݥvqb$suH6fE掠G!Y2y`sI3dɩMB .GhϹ44]^ߏq!79aX<'4P՘?.¹3wICv n 2 8觭,9lswj>dIe٢l e]lWUWh|3jA?WeF{\1&E1Iٙ6nC0*wj23?Mȅoҙ&xt,s}-= 9X2zgWryy' bäťkb못YDjo.bgEZA=NhAXþհ+tPa`&ŘJC}{>-F:ڙ#[;EՎ͝x//v K|87vWSLS wr wUwPsՋz33 ZS;hBz8dr W;@6 9LĐ29XAk #]_ s+jK<)5COp|z^d{ӗ;0ϓXimV [̵~,neQo7x;o] j4(dp9mڭy>8f]}?viW6}]!o^^nzKE*Lٿ/4wT$Зp<*(b_%w2Ŏw l%::RLݑ)paGQ꼾VÀ#gB\:$UXM\ 2p'uGaTX|?lnEmviJUR4O/.2X+S$8r4??Bkf6 _8&'wP**d OTΪٵŸr/՝Oogl1Y0_8U3K-~ h60jI֑_=E0uu0ЇѸQ0i%Dwc.5Ǩl66j۳*ZHHcٸ Xbe5 ߪvgOݧߞ_?|'of`\Ff[VϏ?n>tafhC3S\v-~7եC;p3Z _~; oYJz3|12]Tc*ĝFvT! `@ۺ̦avH/x%0cJo4)fc?@ vO6+à z#,&k/dT@:鹵FaH N axCE:-&) TNs49/E}Z! BH׊*!r^`id(s;-,Vhq&T'P{u38Oފ: K=+ pPXc)@_(i/W~ DJ2U..Q*r%%w$-"ro+;~ؾFµ #{119q|Ѡtrs?Gjdm¶E ֋fbt*D_:s\jR)uUhd9}Glsjϙ&V\9iLS6+Svʆ'PsOM.S`9ɂLFlPFE,4'p` 84Tdbm.tBI JF+T4hiB$ Lc{Yft9̤Ci[Ju eȡU9sg=*[⣶-HiKwT r^`pf#3DFǜ`GGopoe$wM^ΎCqʌH!*HY0Al5smX R`ԩ F8 %B"hQ\1 A*!lhڱuFΆrj=57ISHB-1Ry)W!p4K8RV`>bkX/]>ނkE~c\ !yɰ\jERĵ VszF)1Rq|:Zn:OZ װI*-"G0A)gJ((P2NKc GFqUdg)HDc?o":<9#;pNdAwua[cR;ĬW0 jd>9HysQ bmC 8S B{_g>|2׆,V(9 l 풭1EC blI'-]mz7GQ *J&j{>NvUEk_szo-8g]HH@h5HغIz֓J&4$ 4KpOaZ/(}]û*65'b\1\'[9?eIOk8ۑ3?|rǓXU=LAK@YuVR/3W75v %j Ɍ~ -‰ꦋmUCVE geEhʧej2#r2x}wndW_saY"]v)̙*̐<47>[F!Gj0jRsY9&2Bs2,H4EЦ./oo-1#=?V3#/8'/LG eZ30{-#cԀ.RH\gUiU9]ioZCB#N= *(Ёm*ab8iJmKwtMR:i,@(vhCvXPNbX3$R `4yyir'ɮmsv'^~4ܖD[+X4wONKun6;{JNָhX%epS] d>Z7 )U鄙XP1g&p]Xb{Pfg^͏f"0pa>U ᠘:TȀ'ZK(Iz yI=R:BDN.tʾ" \HcpGa9AQN p 2G Ҙ.$ނX/c" #$$8XG!1z ",iP%LR#1s:օ,1KZ/il|m1 /Pujj`ϽZ !$=eKC^/:KEǸO"6Ҳ[09sH[G“$VV؈W~>G:p) yå O=cF@1gz4‚SQDWޞz&=C&G0QZo!3K|=mAXFbIvxz"ԸY04\`D$| uv!ÅV]@7z+ ċ0hvB(] J^il#HiԠsʼ,JHyYlRb6a`QNiyb[Wcॖ90O[lJkH/åGJ"qZ`MvVyTi88|#=nVΙhef7L]Yn݌ȵ M\k3 H~zXAaͭ׮f9!ad}$zi-g7O^zb湑a4ޙ=6#ܿ=X&X`T9F$㑅H>HԔ1тFQ4`pHnlOw[=õ-W4nWŠ_LE-ǻ+hӘG"h@QAsglr R-:Jѕʬ$* cS鵣7_LT>FZJۯK$ҰFWfoѓL9rq<>FXհOK.`<V8r˼ތ\w#̱|<# $gN)PfWsbKa$"0 thڦ\w93v֚D< sM`ʹl g- lJ[s&bS#œxk,à5EFS9~U۩IҠѰ+4I_-͕b<_fEQ &oo-rӻml>]_}R>>WցPZb~z,F.]^! ̜.䦓Ye6W`Ho.(1DN5 6"v40FJ$Qč+aY->vx^\o>oQJ o,ޣU3$]] {a몃.)]&Lx¸$W/]TwׯmԐҎ>'+s) h$SuMbF,RWE֥K92;9Nr M%]Ë/yL,c05%a𳨢+&9NxQ̹9hvjà\uY>Fa>i\w{& 2ќ9sCɑX[U8䃩OoX(4{mQŞS ӺMkx6Jã;'-#BDlňZ2܇[ E k09?#tjzUv}/; -HBioa$՚gd*()&=  `+xP{@ē@y{@Z=^?K|>R}UOn}\-9Ak9c>Q"iLwNБ8Aeg= q"ctu~|fA˒9 ٵ)UIJ{ݼ:^hcLTm3nX \s6}Ql0:Ao=[xjt6ݚ>Y~\T^f|Q}q FQ6[+7de\m͸LxaӥPӝkttxumY3a`*&K>.f]kQxWߵ~Ȧ6=]j I7>\>'8NX9@dͤxz/ۏϾݧߞ>}3go V`\TDf&g@3xp[[SŶ5m>u[ܗ6+ćϖth{njYK짛/w¥á`va, l&xŽ_A̯Jo9(2UFvWT! `@S\xڿnۣl>KYl L3)'MʰPn=\𳦍?Gfoy4H9_; @:鹵FaH N axCE:-) 0TuNݾ#~KQzHYiO:DVT A#$LRK##Di,f鵂ŠF6Zل:xuQK=MZĀ]:t(NjVq\2 "-HI杶Ő3JE䎤@)"kF[SKyf_dK?ΥȹefNم e>FaP33ҊotxEY7U@p= =)<>O>~]ŵE5X'Vۉaڪ>t1#ںIbv`0οxM8C`RGSJmΌ6>0aLv> =#Mb]=1F[cR;ĬW逰tb~sVk/;^su>c}4k i/D1SyvfMߗ̆?tF 1K6e($4z҂oӫ .ûACl;5uVvÏޓ씿5v0Cc S TU?Kz6Q I p[XR J_xWnr8 g)>s=p5?`-vNC`_@ҋQx=/?|(Q|݂>N0=, NN>\dMmd82Oz"g?O>*;dՕ_7A}EW2|g|3n}l~!zY4m#tz7fÛє+_W/flWu'OVD)hy>"%xj˜9r cɠaٺ}MToPg|9~,g# yQ sϷs摎؁m ot[v AHRVa S띱Vc&ye4zl5p݅Dv?V+9xv"ЈSO8#<*+!(tF)JEXNf)&Ԗ(OPU[c4# !&ho <9.)7f-g*K5k7l?~,Ij,h-R܀~V>RL@+GoW12#XP1g&pmXb;zD[iCMh \,B6+ 7͋?~xMZw:=U420} J^B:`^el.DjC,+ȅ4*l0w4H&$ڐe c}9Aӆ=u2+10BB23uS/IN Y$53GcmXCN:Iۥ@([ԫGhJU݃Pt:\+-ܻԭչZ~;O[DSx5:I o]I)b%}:EP~8D˛ =]H-ʨkm$WE8RX,,IgFJ俧jɒF7˭3j6UXQY Vr{k9Z|L`AпI*{edry˘<8CĂ8MWk9V2FrꗫpsW'ƅ[ߘ5b_TJH" cm2Z).QN8 ((fP?QUs;\f4UR)([E5B3$:s;uYBR帽g1ک9\pe)ÉX!7JS[2iK gilRF)uVtg_ Bg&*>YPii0IBvH%5^Ŝ)(/eAU.U}oSS r_5Ȥ<TZŽI&Ϝ1s  :Z_'-= kbi Ci Z`,d%Є r d18ldGf8ҿ͉7Rw=Gi!]8jIX-xch L !Dg,RQ,uAJRTlb))򩎕oz*\dJcI9JPg9M9c92sLѝzؙhc18.5llB&d7@Lj.\KvzSꚫj4&6SY"QVn 'yNhƟ^֢2<|yr xK(l d\ ºi6S]X8 3k {/JS_^[#\t||UQ{m]2F|=yxuwtq)=rVhBŭӭtLq;tL&$oWGSY t@XG Lz.dTa*TY.\=QW(~:(l62S}8Yb>2l`&YURd)+9MQiv+n  ҆A8 wݺcSlxLfBeT" B0\>]Wd" jĢ+$MF/U,Zy`Q)pݞ6m^R`s@9h WBXk/;RGc<`Lu[`_O,΋ ΞMlA8nE[3 uw(ڸfܻۜQlA;3bߐ [[7_ZIp[:ٞ,q4%Ӣ:wix[zky> Mnv枟y9SYl;qtp0y=ސ|2ǟ65e C mZL{Tls)xnu^: kk\]ZSu x\k-8wA0T[bDќ hdu+YHƓz~4mn"*d3>&EVWqHUƹLmg40C]vO華GOs8͈tv05,z!if 8&07y>2ΪdS:R)Α 8}b>P8JӚspm9NJA\̪FQNO]Lk׷Ǹ'ޞ6Hm-)׽K^zO'ϔ 1,>a󤃇^Cں*t]>2O+ 53}5˔s/3ƓTrDW OR N%e$$6D &A3_=IO,(<< T`WDX܆m1)dO7_GvlQ;aM'(\Ҿ' E X0S]qW]>ӇI=f[谌N`./CTݻ_ (N\8J;0ԋF2(Y4tFӻ zp=n"`P} ^Y* m.bm'dg%{GcѨ5kC=?uɥ?5x7kC1鋚M?3x^ϽIqb{vsj޷h&hxVu  ,Z~b%7Nt˜ߒ /%TĽNbݑ9Hr5tHg@AJβ$tLBрLIa"ZuθpXXW/[q^װ1]o>s~Ϗ' ;fWM97v2UFhmlgz"cKZy<;@x)KyvHN/#QXީJm>_x_Zz=&0YʥxK!HB`C4# ҲN~1IAL#+L0lZpA .S_JJH10xG:gꝃ;LiwLU> lq^:رxƟo0we_ΆwN L H_ȏ:5pHYnbTRUDy[t4Hb.ǪbݝU7]7˔f|%9yڝIiv^&yAy3f55S15S15S15S15S15S1 1 (U9U9U>T*Ss|N4nR1*aTPW¨JU jJU *aԛFZRVVb2׊Z_JZ֞}Jɨk1sMU o^}JW *^%Ыz@UJW *轵P8@ee諼 w }2սW*C_e }]zQk*C_$'Y*C_e }wk1Z)jJyU)*UWR^T+sl]:FC29zjA3b&kGz0{s){(>C{ʲw%7@ΝX)gdBd< A\y(lݺ-}TfM=(e߮)ޯʺ?Fl1c9(X)ar=!?ө$F " :U;QqzxƭuVG ~/]VwnƵ]no7~ fu:]'r4J"P?!wv  ]yb1Jq+DF&-qb'@yb9a cR(s, _l!|z4ݙXwQvcW6^__-sTM՛! b2,rj-z2٨UIp%]IhL{&! xv\ZC1UA?s>1iTMnZb{-RlxӚ<ѝE|\|U3WhfU̪S%Ϻj:bU8R<N0 Ӛ)SMԮa.pBvdmXN~oVu;Un)^WH[̄/n͋Gt|`o1O5[[[wL|w'N EEj|rkh'!ga=\d<ݷI]xt u tR׾ ͷ7W .%5ra(j\wN:\?L}K{έȹ@i.QԁjVy7~ 4tyE|o)\o4w%ٟ]_?{ǭd_.Yd,0r \D_=.gFiIڑdXtO)h;ǢJ|0ߘ} Oo5BH=xyFa;Z=[ާaÁ^>xvmF iy7k;W5L&`h{? d^ǽSԽHGGbyݱa0LatnXQ`b{;'kE]]rslZ]>&FW1tþQ0~CW'fՎ (6\)}Elˋ-gy5oo}^w<շ/^˾|/7h3jL[7?Cjh_chSJ)oyd cT=$|Y2|WY澖y~qXNg1Y񢓮{eV_peW,.1HRYvgw!q306/||ݗ(WS^#GNl9['Sf.p&O/a}[ƩʒhT6E-J`r?| ;ŦTU]Ug M 2M*(c3T&u烫>͐zm\Ĉ %jR.CM֢ʹQNq&"$Pj;lF1wEnnMVUV mV/t,FFj.tVwSoeZNLZoU:*V #&);g1qRs>RSUV[e $H;7峨)L.1!&sM=2S8XYM<8Ś'|v=3\\m[rR0T%^W yCsM%tc}JFdZėE[oڷ,+EB`ji#/4AI%!j(\B+G[gǰOg[F֍oؒ|Ky빓y%oImfk8J11S0HqU6E_l6rV!Q&Z4^"r^WivАe\κYa%Mj\F`&ĹݿY;L%[hܓyfTm^* r*sS|t䋲J+)vQRu)]}p _5p@ȋg*d}(NDQĒ֚:fI-՚j壝m͓.L2G*kBpȅYT`9ZERs JTCm6 pܫŲlY,u´;C^yJo*Ts1Y'j> zt3"Y6.nr7η~Q}ɱS6RTFbTJ}cRe5)")ޕHP[=N<9/1;?IXsyjp|8K'P2^v.HQhbC)j*^˂hЫz4=E5ᡩ1&97yV.5P{kܸ;tl^se]_6bzVJ18 *TE`*&UBqJ)sWMK1I ;DDNk.>_^׳W@N0>Yh7;y=Pf~JP߽s﷽FhWKz l6k kgcJ;slFԆ@NSiiҿ|NfЭk?( Է] пhe%o?xA |C?8G}o?J,5/O}s;iҮ,;GA2 $ LegCxGS%zŊW< 6yw9˸/|/Q΢/ŀ +;/g?.?tlſ\ibo=A ヽ%~f~nxl!^|{H=.O`SQ]VyGo Ӯ+wZm$$ ZKb2R JRodZR &\/ꛬʥڿV|SO?ɱ~ޖ(6BY)ڲMjsͅjcJ" 0CTIOD[}4rprL[꿕u\F x%y%Ĺ9·{j<8g ɍj\=އ/ {,.i7n6y4CK8^3J=?\H?Pt=VJKIg)͐*,Y/ ^Pb-ZG8r&Y J(z5T~oG}uezyypZuu!a>9Wy}gqxŀƐX2)i;.xoJB-צ)0BŁmPm?ڰ..R֥pڰ.mmgXf = ݃.AwF>£7$W+udeVYc}1}Ί;]6uz%ekãw&8vVYuYAnobF12h2"b|`q\x.DfR)x{+WV;$1=[,)ԦR`l7( 5{#@Q_HΤ\1Fenf5?-'>EsǪ4#s_PchvOPfi1JĔhD.;TͳI$)!5듣H>.HwpױOY=Yqk+nR+!7{BF]irRT0٠SAY7_JAUBR ނjcJEbtyU\7˷jp06+fAqB}fwggJj!(`b zOlPoڻzf'-+3lyuH )Z')qȮm)>չp}WeNcmZΩP#z/*[x3(nCdp6 ˄EV'eSQ_< mE6Yrfe53Dv ,}\X@W[?~6pzq2W,S4B4Xh;K1;~d<Ťy䆶益Cޅ\CCݒi!E7S^",Ck4Jr-h'R; r@z-l Iٗ;&,w&NmDxbLђo-r| St2%ޣϱ?.,L  3%2Dir9DPd  Ho\{+X]բayh$j** @`gUS!HƆt"b8[5ZV0]/ 餚:k4Qydl4؉75Pʥp#ۗފ$va`>,w i,򔸖I IMHJJvnV1"3 ~/8"(h$ĿnHp N@e +X2AhR)'UCtZy;k=^y^[q2 묚t۴h>Wī*]x0n(.f`tOEQb0chYx*ð^=$A*FNr3cL5+#?ʁfMa+l]Á xpuw!A4|(.n8Мy 6QB>+0w!R%,@B*J](q'r HhtOgd}m^MU:* 'R}!h&Sf޹T<9spY&r*My0V `ra+};״(B᠓B,Q5qQ0T3|@ՖoKaQBxLޣي9`NaT怍BD@e?R;ئQPˌGx{&0/9IWU*TQh ܵްՂAoeK*:^U_7TR L.ĠA}0Y+ȴw0Ddr"Aj| ^JOC"SDzs|Ȓa840j"A),(X22 q7)9[`~6< 8\4Q*tCBru``0m cܬ4xn$uA%Z=X t 4!WfuwE N& x/T10jzB_e t?kKe$G"$xbܼf_>O$ ^AdxR N8?>mNuj=%XJHn1Uj9=*~6׿qܠDЭHMo~7Թ/ qZT'%9o޾0ok?Iz͐A"+R-:QjdYd]P:SeCUB N]6tU]+O]%ųL 5A_.|Ɨs >O֐LߊFp`$( `#G DR,xl@B^7GWY؜\P]'@2E]K62Zr7uP!]ʅ$l7Axs`7 r uP&AGH | q)Z pG#7ѕfe~fU'!eyu'6 1x{:tk'=<{yppczЛzx/= D1[-}ft3,.?G|tWЅ ՠlhq> 4Q ISQ@pB-O{DZJU-7q -@5Axvvkկ F '+&A~̫MR+YWC`(ehhgͺsuv0n.MTt Am"polIZ( >xLڠS)TjdK4&%?wGg0Փd$KL3HA<8%^oFV1J)[cфPԊ+̓0OZlr4JrwLw1lgv\΂<=L G[bi CL-uUɢPrQUZ 0(g%9Z1ӛNQ`ߘ/{9.ZBSiZ[ Ri*ϟ>P=FDŽi|JE3:S er]2,ƘD䁱,*Q)Bkk"}g3=y)ltcs;<"TJ'^4B"h6|l*g%:slFhgK0FKf3a[y7X:B,mǮrQ: yv鲮~n#-&V@?|ڜ.Rхrz2[^.4Rqԅ~Іiݾ6nC_ !" dyH@#_)oQR|IKiz{Lyf{ :Djwy>GR`zyv3r>ޟΉQ[3J/๎Qfp mp_ZڗnmL]۲q~kqtsY'P ~eMᾊݿSOKޝm~ÏїaOXǎ|ۑR[8>?Ӎ-g .,B+I'k1Ԫ0ǸxóЂ ptqu9*=nd9u\~jͼsY.Z9{o=yߎ@ۥ'J ΢? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ^Z6Ns t8]Ns usmUAk\C6A[q?\_ħgBJ.Z-w6MSHIIU-;eDcbKc z0sqzmǵ |6GX`BՂ1&m~}zkܑGzN([=guM:괹@Gn,7zWg{(L>GU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]EU]}^k{JPjNGv <U|KPvʮBSJ7]]x*oR?UAdQaULB^ /-ʰ Ga#Fz>29D#ZlBUn$`fDz.ɔKaJBvdLCj%ݙc'xDŽ٥yd~.*UJgkٗ<<}5zʝ޻o wכpNg˰G_w7hG5_>Opaњ5jǤx@Fjm49Xk2ꢂI>ǨDL:Dh%碮#ʱMB-wß;^4wH^޻y,oޙw8W;9patγp7z#'wd} ?4ྲྀAɄq)UՃZJ +VۧXMuպjx<9l`i:e~  9Th@: - 3FEpZ l+hܝ;Jt>T%.˽u5 vb=` ӟi)gI2JR{ƪP?7FޓqdWHKuEɬIA1:^S"RydSŦ()VDu^WA!gBG:]j筪i$8qpYU6c5k/׵1ZJY{޽ }oCAmZw1vdtY k-_uF*)O'@:c*!<Wyڟ9e0 ,~0[? xAѨ $ TYI䦖dU@vXR;Ԕ:g*U|(yJlL m'T"+"q`rcP<*~:.j~ţE;ڍv)lM;}w:%F+H(Çj<ME58t$٨%rQ_%Re•ϪRtK0uFpKv"65mT*ݬR=]ҶCf#w@{46ړ5GtХ uj(-*ծTje9z\m^vC(%᥽3l[h78\0oRYXzCFJ+T>QHUbil_Tl}&_l >CNoN&.2j{Cqÿ& KZl/L HǨNV?&-qwvrOӸSe\A0 a0,B-tuc*DDq*4q o Tzu!v( {Ԯ澿&Egs ;]LӳA7f './a GlG^$V-ôgc&7Gwco ڢ-4:ҴhYzlf07#sx&.Y<ՆOGwUu h_վf+?=^/x6)\(Eˉ#Ig@]i3)PR]#p^bfI9"]E- xv f !5O_:Í/=()Ciy^W\r A%j_Me4kW?EK4:[*r(8P%ۙ?2T_JS}XӾh[Ϊrr97JtρZ:DL)e>שDCn(*z${AwtyB[L0RX}53sώzhEV i%$%O$'+-= gx]OYz :z< IJԵ^(nMAMh@@pZD@A=" ]DaRk(c$̸p{c89"0i5y Mђ(:>E m9q]3$}4 GBn"hn}{&"cG7g-r1,H(ד֖E!OTBJ r%4fre\P%p/Tap9FkI=mP&b`ZZ6LP N$c hJ5R8$P\f B}ݗkB=!Lk㬏}ĺ +\ >* B me;粓/DNa}kA @^*ͼ`ޅyP& 9ӥS{R`xFUq"w4{0ZaGT+)Gn<3?:O V{o8 9XDX x ''`8SWǑ{7O\]Jf+k˗Ǔ*jJMo<"ZZ)/v\x".NUo~׻Fѽf?kwݯ;o^N.^TVc^3?Lr{ "DOΗvWhI3MZԶN*o # G5̫%h8jz|=g?YrYY?kmR| %F!?r+Ghz*}X8 Ƿ'߿oOooN 'oQW`!nG`ܚ_t55L͍`j|قo0]r˼+G^/um\OC':]LYz{ d~6$5T[J u͑YYA aj@_xޗ~SzZiWuLUzݩas(ֿ@} v{OƾJ£N`@Hb VFb傶Vgq4:+B?)=oNG=HfDS PEҹĂKoqssuOisU!3,BFR:fK/=a|S,V ySugyBZ)$*|bz*TsUce`K)idž/,d)RL՞K"ً)Y$;"YFYL%G ]0==7Fߠn0cpƨ/IQw{=KSܯMz`ǟښv79X!4ꔈo:RbȖ5jRrJP*C%wܡ]rJx@*CKP*C%wܡ;TrJP*C%wܡ;TrQWrJP*C%wx;㭬%wܡ;TrJP*C%wܡ;TrJ!ٲFu,JEs(OEZqOd% w*w1 <}m)A<)"h^ g8QFg%ûZ'qQa:cArR 29X[B,cJ)pJs*> C`tzfO+4GNLMעEB^Db ׎RGB`PbiDPr N>r&_ :;$e\8szFAh 8P@NGit*Baː796ZEGSs v*3ߧ<_o0U7Wz(:2oqzyzZMrviLlTsw^L./)}^CoCz94ثk0ectz+sP/FZL ~s:Mq }w:Dڴ~ s9wl"WYUYS] K׊֤lJNYHo=$0'IP!/ wCi GcBѰ`)xk"&DHFE+=Ӵ{%R76*m긼+rk⡗m = {uO#FQw۸5. m[?ݛ߇O+1þz{Z7mVg}F>m*Ec06ڲׂ9ݷova[ R-dԟ]WTʻA*d5DU -0kG݊J]@MecU|TV:UR}æbܵqw|# Pd>?jfS]fpMv\kdB#xC-@rR!Db:Q"u;**uv.ayiw=)y4Lv>I}LP #db&ARIY:9$\ɓ$Im=C31"Jwz, ro( 0q,i@\L3.jT9%lglp,{U-r/\uW{v8 Eӄ8;ME`>.h@PZBiۤ-<\%j=[{^׬n_v m.I0L6$O_ڒG}%[r[֣- 7.USbE~|GqW1_]r06{{7X=䳋S~h wb!](Ϙ+duEШ;ó?jMT.%=jFg80NW&C?ܦ#d6uPV3 0Urg7wFH3ҩă?&8{D:@|CVw|X`[`1T;Yj`WDvc {a'gY_u: δT04Kqj<1QM I (% \sƽoVI{ Iިv NaM BTx7dQ.~ &_e2 <% 5b+ ˄!A젝/y`. f3HP@ijyHuOl RfyP֜,'NK.9lf$TsHW;KL/l HTaJI81(' [9Nb`bYAov#oFS+r*MG'(Ʌ%7o㢿Dkbx{g u?zgjH8jjh⟆[h*Q͡.H)D&iK @cC RRJ1"Et<$rAWmNZ[]LF?͉ :Tq9|(0QA1 X Ǒ)K,0k@#n(4 ~ +I~8Jnk6Dm:o ^~8!nkiG!fE+x sZG5eX(yƮqr\ShxNeVkأZ_~pcU7!fsՈ)tw+  =KsqQ0]8%$ND1P1P3t͖r(Yzȃa %*؈&ns"G2%'8&.UyVyQaW ˯ogº6fu;ΏBm" wz Q >Al]?uM7t:oylVƖ }c͝7Szn\xMoGc =1Ϧ ;oƣnFpϚvT\Ftg6{-C%j|\@"*6bL3|2I\O2I+3B&)[ #aTƍ)Q47P1rMG_ p(,PbXa e娱Q#=FzGXY0bm0Dz! dNJ&3å0$.;L!sJXPIĉ:1+5^Kt[(Pe>j&NǣƉĭWr{'C#aWMzz`w6c}ݻSpF(X'Ef. bxW.irkZUU(B\)'TKRGi}(&DG Mg;NaŁw'k_kA>ZWQI$7c$/p#zTBΘl2 A dJXW<1)xC"g]rǭ yu<y߰ӤYg:sš^^9f}Y8Nv2=ՔJz PXhu@ω\SM?'.?f̷|=b(87[ʲ:%NV;K+N}B"Qy eɤ#m Ĩ f1,/1zyQԣ9{aԿ )˲E؆7[,h0TNsQ ?>eX%TL#> -f1=\:v4wF{65_nG`o`.Aס.%YOZJfM 35Z1Y FhDl:̇$yb\VNT~4*R^a&E)֭WyHrߜ(| ʊJ$?'ɸށ(뉾23K+YSĬLl|7?}?)ar& iltBݥ_Pnf`gc@ t@ FkXnvzRj)^uy;!n>q՚I&Ҷ5vr˱_fMU*@mxz9p#ERIҋ:Л5YR;DŽr\uu% _imC[m;;V}d< ߆o!Q"P5=[t%l"tY1C\KG><1 @,,*,}_^*;o5.]dv*m}鈻1L<mώ$Qڔ n#>  )~RJ!K 豼 K4'AKMKF뱋(,6 5>kBp0p>ŕ JU)3PghMDBjTPy4 Ĥl(88Km)fpk#M lt(HsDiyUvɗ=,tbtsxǛ|w_qs,YA RލFG/;H< ^hDk]QnUzz,r e{CcD肞Ct`+Mms7o3 ׮jjmOԆA*Rs0R$;?Z/_+ݭbC脌wHr P<0VRцy)F0$sJ&_㫫o$3{aQG3{aqf0M !o ̫!n:pwԚYXe[8SYAi4gW/0ΎV&ծMO/ߧykVib7nnp>@\JB A0D0g:6HĨw)6KEY4%)*?L#;T_VVlnDqyAVg١?V5 \b0a0"b+UbFx0Tl.?ݘVt+ǶMBhK fZO̴݅qɦFϻ_4[8/n;QWr6>gL4jEh+ O[gy3X˘j, f+&i 66=i&rZuky̓fqN^ڬ9IA_ۢff{h^]x0 CQơ?tpl qWZ%SPpUHPSk!GG1 \C HrCeֻ9-1SlUW-nj}QG-D )yuI_jKTJH:~=82jxoQ!*ĂTUhDq PP/$wP.K_nJo/A䴷`gy.D&踘`Dߧ'2{K?;rJC^Jœ6(rDŴw˝QVK%5o'9qu>r %CS'׆*aq0;dV)8ORAǁH6pZwg5mj_"=nhq4?F5m~۝%+ *&eW F.*8$Q61ecwT`첆g_75~PAA4q'I0~BOqz tLR4=rCQwՎ9;l:{lfrd4e|oJϊF*=q8Dm|OV|L./@ў_n=]NQBNj.GZ.5H]#u_b;fFf˪.ΖbC  3[c?ݽp1fpł\SUObG:_Qv/]~*#i}\Ja z[=Ju* FLDO`%!#3׋EȾƵ;] rQ3Dq5)s6_o[i¸R{H?]*;-;9_N;eߕqg܉UiEF-*ݲ[,r g{0, 0F F'sťcVcRr1\rWQJru**Iر(7H:!wE('TJy}UBYtW6$0'㮒RRt`ޠbTl?rL(Kkx|;ӏ/_]6ʹbO_ UƢ@|O3,,0k@@R3B"ϵ:æAow:~\2b.4by@qXc'V@C,-|cBohZOYt?_la℠ g-yVI7$)_qg*_͓8]@!&HE5;oWLZSfuޝaT8DyPQ,RH 5D2,8vToF)bNҦf~f3zBE1xٝcJ`Kn#2#;P6Ilǀ0Ab&EŚw~A@}P,%F\᯷V#Z;kOFWaC`Y/JJI.%?T*Wb Пh%6RK.d!aIdte6iE"0 "IJIYn(}:ϭ÷汼* i5kSyR*dU60w>'VX^]^/>k :e"2[V1E奲QicfL' % p Qj',4qJ#al.H, zgI p[#gG_|s^~HօJ5|g7n&uYZn|]69xΣFMz=[ɾTaJor&C^;Xd\ĄJp $:] f,MjN:@e畎Y-q1 Xe-9nˏ*% KϗYwY)A 1 `/\f)Hõf$cLB :Q3O6z^jXiF2<?xt NFHWE$ǔŔ5`I##2;Fj;*}V 9_ >JaS[H? 4u+ޓ=Wm'rS] 3dIy ?F/fߋ.͝SMO_\Ed}oɜ,q)#s˜8`G'He<xBb|5vf#)9adm#Wk&7]ф{\rսW Elhf_7-v_z99wDk}wiD6`Y?$]qJ*GPjZѾxY޴yq3ҟ靷/' o bT׳||?9ά3(6rR9|jpZ1n$֎ #_;RyGlp0e`Ţɇfc ];`GmrݨI|qqc!e`MoQ,]_Iz^G7,ISĆN N;{qJOo~xc|wLX_{=i? L0ZjH$v/_ 7?>tm M5CKvZ6z6;mf죯zVwYk}q,Ezl:]Ltq=H|&6?{qEMߖ\KxjA  Xvn h#B2f+`S" 6~ncY۹L,Eub\^e~RD:tG,80 PDG04 HG`A3f '1ȬpYGo\઼Vdu &Aţ&~<|pCcob.j?VV)j3/QuqCL_6B U)bcJIi*gdxbp]gZDKkt jx5:%bťIĹ cރI7DW4&'=|$6^@:L}nvHo*']4KM ֱnlDozw71-'[YCe M (8\]*u~Ɖj,.Sym"wW EA΀ mVQ^FnaF$ QUF)t}T\Ë뻙NK ԁ38'3-" c.8SP\!ϕB.pHf.sx ^c0gIj&1=|mȌgZ&#\lN2M+Ϝ2 gt/u-J-7G$$}ּ|J"lZ!tNqt 2N~9YOFu΃rGE*Pj[['OYvWuԚX),Z|}kҚ5$g+WveS]g,5>?`Ry#ldLI΢`KWYQ%JX崧lتd$5C`N)DF)%rV`1n$pdRmK.7QlkeUʖedװW\؜5.ü=Xu|5L'ڠ᧳mSdVDa1 &CR&F>Ls> Qy#fk&[rZc.Gx?r|ELm?*BQu_Ԕ&2J/`$i\Yo3zuWӀ BH@V~$B-*5ITN|Zk+#]Tˆ~W_oUzoٝ_kA мn h0 W4@/?Q\:K]>!?W/z{WќԋcAzw~sm{8AZ!BF^ ScJxA4YϤQds!ZgL@˖5rvfBۇq-KP Ē,qY̰2MWX`ML3Lʙ)P i{AN;,Ӝj䬳:;i)ۡ"#EI9Ĥ/F 9:Nr)Gxg#8*mckkI۸,"]ԴԐMeV39oW9"0dcQQ2)umyDkYNFOkmܕLTxځ="LjIVaaLBp!$9p@Ise8أ=; +wؖ>G'(gIU7 o r $J\'6J]&KI}8esmZ Nɋ6Qd URP)rUBNQ(etQ+mZH!.t;+[wɯ5Qr\O!o <2#Td@}dK2E3mBaVg֡@K c_x]>]ZMaބe4B%z7M\KAR_柃Uh_Ui{O랭Nz8)gg:0__ lUmYc#ZUӺuPO\eȠd5W!P9esXZz4wH胉c#-waD0N[eQ dM]oIH3ɬщ' LDA]I.p~#w AB>$J3%ׂa22Ilę\>5Ύ"2㭵)*j]QKjzs4NmnͺCjYQ&nޯznYΌz^hy= 5w7ےϼ˚) ux>,N{dKO!e4o$V7ɮn`2\>p (U.(yMTW.XSV tgwvWdutFf 4JGV"r2c:t%9I1+LΖ2 jceDR@ε1`IC=5!ޠx$@#8 j$2pkH ʑ}ܤ5rsJ q1c{O:qڦβܓe[i0&SWg4InzE2!+ul'*hZY:뱳kmH_evH~0^'{6q Q-qM-_pHI%K69]SzrF^^{-/bq.]A$IJ6 HJ Sk y/JX8-zq3F*Ԝ0U`PW/ܨk:*ߡN1*saTz 8u4vLwcJ z3/x?| /QOR̕_`e^KQ''[OC{̄'ҳ7D!:T52 y hF2) KnLՈTQhH$pF`Zs&)K R@ %' 刖yD4^2u5Y׃5?ݜEf >A>} GI)arNeBH{+! 4QPgQq4Lꄷm3 $m60; Z4tfPLmovd08ۘ}3D]2Te)yF-!ZU2cx)w$OvVͮozx.sD?舞g;k麭\ݲBcb++ ۇkۃ3 Cwsia,pXz=KOSv%qZAq"as!M/Dm0DnIrUq<Įx*)i'E<)z!$8uS½M8{>Um4[]Ye{dr()1r[Gp:pC,΄uVp.^;kGFL{?("d^(<6!!Ri1D+%hNY?P$P$P$`Q xAHh,$ށ5te/(ZEJ붷cr2ܣ93ߵU,9(yi8P=\k~<rюБhrӋ\^B E d21`$VKQc'˄ʫ>1/ MÍzhA]ÃAs|5;?UAA*k=,`A(K&WC)ʒtśeTjeyEY,J>i3U "ŧ,=!RX)}P J(Y (RuP_h)TIU),NyGu_Y_:f1aRĽrtV򏉀N>Eޝ'j {;]'n}}=:Yrk?0ŧ^sdr2"v=pϟ]\^M'7{)7V-';H+aliݒeYpÝS6lB'ېw@2;sw9gYwDw,XPlY(nCLYŴ^mJ"Tj۰&p>-i6,͠ e5 `r͹g^T#8$5c+\j4AF}=^AQN܅wa{ΎܓN`e_LP3p$AA{+0{OC- osQT-z컛Mq3a}ovehtnq$js[\s<Uu'ޖ'-F_\wIHKʦ@JuAiMj{a2)G1Q0]aV g }*os`EkVN0-Sdu*ѐ[&!Jgb+y wg=ͻZZ0``v"@ۙ0;2a$N~ﳌ@ӹ`ߖzh-RHJJ@IjOWJx,ث (vYzi  j=^D$%b-#DSF9V* |Ё{HFC"DFp)Q>piԚ:83. 3x%LZjޣuCS$OZ#gMiϮkq0ʏB-1&'4UO1ktkm}wzuZ*[ Q'-B8,!F Ji$pd5\x@wx?(CM6I U L+`ZKІ‰dMF,ZoǚqOa8cD_1:ۤsZO5DkxTH^F8"YbqH۱Nؿ۷$(3  &:b&'@GuPFX/z vԲ =&Rw8\;Vny7S?\82H-UG粓/DNaT ?׵ %^R95^0§<(CuuTſ^wLާk~G|yŜ9=`U.|'As ; GZqNy8rUsTpNW秡 W s;7ŅfpeLwǣ%FZ% ED:p?<"ZZ)v \8C'P?3OԌ.ó[~_y__x{5:k>xw- pqKoӳr{wEhFݙL7e3)t&!tltAiaG U̚|8Lwx8YY=!ͺY グU>`%߆!+#~Lh1@bMa+n'1Yo0\~C?> =yw'޿E'\\g{ T<ߞDO 榭q܈5ͧn|yn̻@|Rm/L5k-@,~vU!rL6ӟ|t3*eЕA,+H6l*BsS%ݑyq0`_xY~S: h-rR!H~;Mf89y7>'G'@ m_:/]upxWRzF{̂ PE\t\zb%ꉷ9PuL!,BFRhalC`ʹzاEuoOp8}06=_%;+o:H$c=v:hIfA8SA O%S*7d:InkS߁Kp'gǐ g"wN[oѨ (R@e_QҪ9׌3e EFS[҄%,C%\4dQ'5r֔@1O;^ƞl%N 0&ȥ(D"N'kDULe;щ@"pOoF>D4V14R-|HS^#)% 1IIVkP+Is!`dG, rQ0+0V$Q !+IcplƱW3 \"u4q95ђcXƐ]&"S ?ޥ">ղMjD!p%G+"I@ӅsaR"M9Vix'S0V;HGkX-K~NЫ*+M 7!e=dڥ$S@"L`6MԚ&4RObϹsĉfc7˘\gbp/y5/C޼?yW  ˛' e}+n=ܭ?~4j,h'Qi( EM")˳՚ziћ#6^G x ;Wzxm,;֙xE.w0U6%74s.~TM/_7W~_ &/jr/wFWj wXcHƕWְGrcquJ𺗫4(^ ~R3ظ=w:@ }ƵS_emc)9ΪҺR L%x-}+Gwa`Vܴ\;Ǎ4ᣱhXqZ0WFEL1Vz#0DeF^dMC_gp[}uՏJPd['*>Ḍ?6ꎊb۸.ߝA`yJ̴gEKlZ7%_n㔿ؑ[hgzWa0T3 s 8&_c-%BrD2yq~2,pBlE.Eٻ޶r$W~Lb,6Oۃ++t`z?̧=[Eo`[m.<9_x׆Gm\tJe}=e5^?#].sx./hKk 0نVޔ6c M o#_s))QhSUD-ǒVlv&jm1VQR5c7rk((g Mg]hF]xP]xGQ{@^O~sqf{jیWsuu9q -yRWQ7>QF\\ H8**\WmTmh_̠LJ`M<;L}sOQFW.U95vTv7WkZێZ{D=TUGNd5*M,E@. vJbնL\czl>k(c=@Z)Q bP&QP X ,f7kdH=dlMfQFI;@(iTDžکF?'=&0Q!geLV:R {R/^Hj#^crǀ!RŐ0ڢ』13D|ؓi'-ele+vTi%"Jy.hk" ȵKݗj0F6!WB+Vhq2z{S{Jhi8`ׄ.Vo|NFy:ΨPZΦPJ7jɟzt2X(JkV(f ]_"eVFj&hpp=yl-2* ɦ0T5 ߌ7S4ξ^#CuG,}r;GecŤ1Zj2aPYySX*ݧT)ǽ*%ןI0P]_okklׇO:{m;ez$ԶNz^gL: R-  X88˧l^'W|cg72MJZW]rDvߌUxqvlJUA3bȐl]}{drE'NOwH)]oQݾ^lS&mjNQk^d(gg0n.»DooAv/fa`]ZFr[ (A#ghɾ ̾oq~=}i޸g6+ٗ2jb?tjo URL0*&*acY5?}'$>X~^xd&ݗrwĥR&(8iz${S {<#n5vr/V [ݷ;4^n>l G8)-tX Aҵ`٤k5rlҵ+!'ը\1]It\^|q)M-kv0l!\󐒋;:J`A2bګJhbD2UG[$ WuR)6)CK (l+c 5BȚZHIXS 06ǐ2F9" dgj`H7rCz .?~5Vр˛5QnRfo43?\LPPD݂Ѽ>y{ }15J f/0`6ag>H!0D`9!jNڒ}Zq,,`Sfg*e@o}6 Ȗ?̊#C2D]h?uA-lHxynYzy t6XDn$mŔSl@hm?hW!Crmo}L󷻇7//9mY&eڼAX7o_|\a;7zzW7ےϼ ax.߲pqJVzaEoaǿu|9yNu+H?J6DZ͍VBFk ^FvyQQKCߍc' }mv]?6/}l6G_O&Z }aGWD\xBF< WMLL-1b[ -kR6GCmMcSܷΡ!Fh+j4\b5આ "Hhʱ ]8M9j~{#ebh4$c/vXN-n|?Cyk|*[5Q2|])yÅTNEksg3U9h $v!DH%X"M{D*vnlgqmk^VFG&G6ōk]1:ŽCmHS 8/r5j>uGXGGg`ΧdF]5jw*;ʉ?/^=mMןo˯K\簿2Y\}z}5|=C䁿z_9$r-r&2DE%ysg nȕj-6B4G_((_*4O؜|Msi/^oںg[ =[޹s=WWӾP펿V{2+l}0ha;"#!-0_ ޝOh#|ꯄOQ9M)믜$F Q8Kf]NxҼ7֦Tʵe3l^Agqu{=P{ޓv؉5c&g:1~9/ b傲&Ek}˪.CVb,cM T+!`,,5 [P\N.BVOD`ȍkv;aCbގ)'{c[s8<ĵmG4%ǎZ5|lCYS@o#C}TsMIrPFmUD-ǒVlv&jm1VQR5c7rk((g Mg]hF]xP]`sw'ٷ/[?77 _W?56IE^)7*BArX6u!VQ]a܆LEنf ʤ4γ0{hS*`hѭͲbO5v7rkt=+hnֶܱ`{jjU@F1pYP\;(@Ӊ/bնL\czl>k(c=@Z)Q bP&QP X ,f7kdH=dlMfQFI;@(:>ϙ ݔ2V)Rl֕1ki8yyQȜPK١Cv)E8ʖgnkrj31R2.tAX"Rz̐ 5Ƅ̽kȼ> J 㷃~ nO/?{fn!~N) ArKb@ヲF9MJa XVyQz4#kLv3hZk̡9mkƾ ;o QY=Cz8/@/j.Vn 'yNhƟ^V͡Wqz)8wp:7sD/.'%C<)+>uzuO~yx>Q1b9ù*n–Km!;7Cgl!諓}|˳}>@ ͗ P ɦDnػؘ8뽋kG7$433_LpƹeT `OvuWMCMji2~zu8FϬ~lV Ko} Zyrz(Eۛ|4MF`Գn|M5B"͌rYO]66<+f֧x:rk]r%Jڕm>y7hrutG bQl'HO8 m| BU,~0/KVr,) ]@ZZ6O%+ ^rL-WJl20N0J5/<"mp&6_IN|9 ڽԬɲ ж&́Lˤ{+pܑpq͜e5I@1ǵ#efOJeC92|;(>ܖ,7= [SK,킅V!`l!RtA fV=H a479fc)RvR(S&j8]%6{7/y)TwGTdv± |μ0F8rNX`ԠEVQaK%VQ?mo~避Uae5ct!-JA7BQ'^TRЋSЋJ)0=-DЇ"(7Fk5db1Jq+4VLZ0FO{ {=MKm̱,|1"o[5Q!Lĝ7aî1q6jMb*UkeA*VЗOGkK&~iCu7k֘;1Qvg(ve ^H(@Fw&@Z.I4%` Gd98قQgR?VtK@N!imgheX܌aqa?MKǯ+3sGi `fD%`?Z܇^V4UV1 uK*Ş1ij<0-Yқ^M;(]'ԫԦasׯprEgs!:mN=o4]vxm76lsco$8&:"1^CB:tÜ;duj{̬ܱH5+vm1|&nu%+e8^&5r9{ē&^vJ3u tҢmW=s34w/W'MԞj?1J=||,cڀ,il4QdHxJj2 ` Rkt! v>zCD hf4Y6 *q2 dcAk4c5"9T,ApjA½L4FA>nctIf2^(&%=PI9H^FŸ=sapfPz]\n/-\/hϕxD\QgV%?i].u޺C ;N7+J<Hx—΂9Yʂ3f(t0n٥VB0pn|pYYR7(9(M%j:PmjmčqB潼((GE|ԗ2p1Ŏ*p~wn]Գ3_$i^. M'?쌣ZQŦSKe=)Xȗ{Y _ 0٥aLAG Y,'>Sݛh@襽r lFH`^%&R:|> !^Sxu`.V5(_Ѩ5ƺםнOsOǿ~oe/>o?ΟpF)e"0x|ܾAg]ϧ_7Z&] t QAR79%>i+'k@B|z~[wJIF.13n]eNf;nRIPO;rbAl?kLJPԏ)2ȹѲN%&``'bY`<8Li H .-Rg^2%ߜ?Q>InO<C\tBzjSoqsY֨۷wd?JCVMYh0Fǩ*ɋ hgP^+‰Ym Q6&NV֡/\0iy ئ&=2܆\{m`V o:@7Ȑ!1JDeKiٙ?(;bg ghIm֙3nKr`rJ9Bؐ1:2wl[z%珩.ƯO?NG5z<6{y#Ji<7 ULfiB^E/@XɍNfc#7vԨ("$hTuwA:Z;MW-ڤ7]q);+̈/}8 N41oT#=@ -VF:FL}n⥷$s=>сր=ʌKhltHPs%ZZDB$ (ʍ˒6BL9#C+̀LIa"ǶZ5.]qnrt}YK,ߓ^d=m"78[ A=`XNqNY,^'4exZuRZz\,,e(vfΝ!/~J )KyHN/#Qo [F7t!7RtT^GB"Dvf4$%284q=np\!ehč<ݜCO;0׏rB4J2,!+k e9G:sLUF66߃"GSCg_hCo_ePw%|ۺc䎱Ǯ|\]:CCp< P_3k澇pjLp kXLD]w%O3 \-o g`Im@Fwq}~6.t\(E08:wmm~ؖR}3M$8,e cTDʶ|)")ެ `)4{jzS+6 ?tonN~m@?uZlcE;/Sl'Hm&qA`M;@j.z ӝte% ޓ' h$VK dBJN/Yq޾&^65INoYxn۷S=WmD 8u1By]WEM2ތ C,n O~wܨSi%*S3 Nn$!Zyor֎n]6 _?s*or::d gwV6lΊKJk)zYrkƝ=3&Rq[U2Nx{ZJ#S_nت V@W\b@gs)$L'iVI1wyBI;I$qe`4 \iN:n}Gj%# qo#,1 60l 5 UVry:O)g|iD!%qZ<qnbElqJ&F92J„ u-LmaTqB0UL)j 6D`#6 pU 1G5֚KB(x̂gɻfs\JJ"fSyo@b<L\ gWs0LmL sˍo"*.ݾ-7OTu0qs7pZtMOЯ,H^f^7zp339kE~P\&7]^?B.o&غzҚ[?|i'*vѬ:Ė[Z[7v>ӝ zi^n5MϼyW踎]89l3OLT5'?ii9glF̷9$I_2-I?IgBg[͏͹F쁛p&>gaM  *-1^h4p m^pBNK+s:2 DULNuFZj/61fg1Vұ;rN%(0jIpQ̜n ͎2!Բ,\щT_tY>x"ֺb~jo uHK2!T(۩+&SV|M"V63܆_C(.@Dj:c'WHVLZ0PEQA8tQA8XQ#*:f)L4wIƊ餘0:)%vy1 >Km̱,|![5Q!Lĝ7abט85ZLzBp?Jurֵr^)ʰ̓1{?]7Wuy\[2w$2gaCG[T$mRΦI XR?VD[ 9R{Ccm)JA,*AzdCr7yU[^TB͸/ ,5Ki @hd⃏t.r;%|h:1"3C$r"> _/n0.5i>͇nc[ ,]KKկO6[5nNn,Tt4X#FhJ+JXg3 )onȢw^$.oqX xpY[[eZ|%*LyF:ۍ\J92[l x[ ]aW/lXQ(CMME|RрzKW<]-%Ⱥdӟ_|h(im'pcBUP Ԟ ٠fctne)-My~yNa\#8 e/X.БEk~zh9D9D˕1[KZ]N0J5/EPLl$xLr|p5I+0;`m=.H|J7Na&z_l$ՅV6TbE-Nag$IS`+gRbv@uK(rV,Jd)k & $)8T> OJBTF 'U>iԚ4q@xcxgBȤbj\=lL TϿ"3/$]$X+IVЪI(巘dIRyOu= XZztwNu6gL(CPF %,RnOB!$:~@BnkLNTfbZ%96\WP X͘FH !)!hPOp_kB^~;ׇ?_cDctIf2^(&%9PI9.H^F\q [jqHzm(?V4mbH*3g.q\?X0!Od88 FX/Dm aC&w8Xƥ8l"?~ όυAu15|8O2:OՆ{6\.^0‡Ҩƅv;cJ}hG)'(,x`@qOQQ~qC/.pt{NgI}m۔J.r.PK*/uڤjmč8 NQꏊwmbOt߹qTR_^z9@5*0<4Nz`ji%L6uk.!խ;͈VR]ntUf-xkϽxd}G*ZQsӡkt՝w^u6̪[4Øj,XVAw?=њWޕw<䪻zV#$]e!M,XspJwCrx_ w fSENĿǝu˝нBu}ӛ|/{IJpF)<۫D`Inwѭ~[&65& +uKG> uEhaV$ĻwPNOb̫ݟf"+U~WPl4qx[w uT\GN ױB "`*Oe7k\5X1[Sdshy;Ue2DZ׫QK*; #W`=1֬͠xG7rxE4"3+cRڷbm#-O"Թ3h7rKEP\j ׬uQ'um|v1ȹhq59b9$R^ >YLt^rЋ}XMS0uIWeVokrѝak' 6k#sʊ4GP|͂'5kW$_S%ь%xg\͂#)8b/6 GvѦ ƀ.k kD*gQWV.X a(AU|rVՒbzr.tg9FΖrQH;u ܓ8o1KN*z*P%4 o @ɕZPE'sE~uVQ)7XXR N`5Zdy9a$H֪# eZݠV>vh'd搋Q k(lfnd,CF8*YO|%"!U6[+N+| &A7ÜsY%D- jGiZa)^uc]uudAZcA6R !VŚ+k9ģx 06oE:qȮ;KNз˶8ve-ᖋF_A$Z8: a X}UC+d(e_.dgYת K>~(AF4Bq@Elr%WM (}rkMhBoBC7W|%3A'[ŋɩ6#Pdr2Ƀ7-jm~]8~wG_:Dsu+ln-FBo#z$Z?J II jj<ԠFqԨ~L zAN2jܼ]Vǃ-2AFp,s6[^kElrYm7mm*=`;e5~9y(kUm*X] R }Ha鮇lѪ5@'BQCfIgS2f6DWj"n@k7r8t}ll|xrgGdW=^/Gӓ#,|OB OP>Il:^Y{C%JZyzb*EWȳ%pUC6ĘMAjkn](j=' Gqk)5% eR5*Q=)ȅjmušњXEI֌ȹ_3ލREv]u ?.v!z|uqw lO͓a+تJybא  9z`@"`s]5&]LCbQSRZ5$gW*Pʄ]3WȞ:knܯ\1.Ekw]G{gotCѠe&./Y̷}K. Lu>DXȂ ш) )VGn=(!Rm %b{Tw#~}8µM1FF5h*8olek4<"&\m* ɰcbQb*uhJo8l;hW(i}ِ<׃*~:ww"h{}WUvn=}_1,!`©E0MZ'':%nYK^-[Ɩt$JVkcK8*l)*s x3Xa]:ohŬܓ-c-\lʾ^Su`Tk[ڒ3$fWEZX-p_^4${AܧP¸C!mbk> 5a.'&Q1*Z c ,(ِ- M5Z˳EKmM[U (,3(v":EY~~ M6` Yv9춆׎[^-w>?W|.hHC !v6/+g7皧D^ȕOXt/ ba&f7 wO;]WS ,y<64&7[Hs$@sOBAaKFrɯhy*j'9hjM.fN[XPrKr˙L,\/bt~e' ~?g7qP%$3_t7<;}_M1 ΐj 67ɀb \ VW%ZMݓ(u^]qWx~B is9˯٢g/_fV˙(F[t(U\\LM#9q$coaa \eE /%3fNjzc^Nx9Q=%7]I[.)% l}~|ۣ.E{I?#!nhu6[.O+a?_h/?_}o_wYfs+ò4&"@xCZZ[l1=|·a[^Nyȋu wl[ :2h8iN{Xų^IYE1*20[N։UI5R#eP!$b7/C;見Hױėft505=ubЦ 'k0/0}9p7[xt<;o<U-nlV~PRX|~Y]eFlEX-2;RhKb* 5L UlmAoR [ `'&3V+gI!5\V5Yun/ranqS٦ o5xwCZuE{N?A2ة-H3^RF0 Pj fk7*[OF>"1^Z+?T'SC:NS }>S_f<409tY ނ)~ROzkY!$ơ"ѳl3&Y"hkʣk}$xg\VHEC lؗj8c6E_l6t]s&EQ{ Ȍ-^9h P%fOΥV0r֍[=,(FRZ5+TTVdS!Z $(R+S4wӏ'+**%vQ ZBjF,/'D)Zu$cPʧN.#Br1 8c S-4L1ҭէl KȡhQ{ލ'*f6"[`Bd` aι"`Lm-BMn,·bB<2hVs,Wtѐ$xVnbEIޕ`Q<۳v8dW?E'eXso$.L[_A$Z: a{ظŠl^2/2ų|ݪK>~(AF4A'[ŋɩ6#Pdr2Ƀ7-jm~]8~ն\mM\.ڭEc>?\P. 0[ uF;͛툴 i9om8,Fp8,!d!4j,FcSB TؿҪֳ-:k0 IJILA~j24Q M±}Z z^Xp;rYk$EtU6hz!ФpT3V+m5A7dx]YoI+I;ȫ?m/L?D^1*n{FI$Yʰ%,%""/#2qp8;W/*cwզ6ғBW z⛻u%B͠iZI)j! WHrmlmډqTBN#-.%$]tڦB0'WJ=B0 ЀTkLq3J9/lB.5ZXk'oKomO~t~ԭ G{lYS" 8n XC{IHjEK e@ȑ ZZczlT0IBJVdO w.1k챛c@.ⵛiǾ^[7zn\ALФpmIh8[d DӚâ@ bёy42K \q \:q8jaVL?8xij\JJ%й"P֓yR2QXQ7kn#/Zh'i#z'87g JbPxb~&d b$&-* ufl+j\}l%E//~.kq֡<)jMM}{ݝ>蓠8 y,!4ȞAfgggNA{v~d{tH] v1uH{5g+jsK'ـ18Nt"Y eRυ{gUIɯ6'k+8g5=*\sCzϖa-4aމV-=OʾOhr!E`N8 J"3KK?eӀ!NfFT587):xpQ"rBq8kO/Sk܎} $W͡)㻸~|?~6;&*r" 2 €^8*t*][_x~ф8Yt2t}v'(ZW3ƭO-V޶y:} 1_,ՌM<kZ giS}3zH[y}A6VIQ M*x悊RZlV}= CtK0dD[0J Rg刨jQP8 :g!ha] U 2d (AZrQ…diHѩ,\&rXoݛ#w ͟?^a7Z?]~ 7˦U٧fh5 86^\rc,% N%gkzkrƪO! hIJCS ^ d/f< k($*sL$zv\{Y ;[&4Pb`'bZKÛ3tM2c݌hOF`\$ Ɍ։̧Yx%4X8!2iQA#!@F]hdH *a 'rT,dKVdѺ&GC% 2+竀kCLc5şw<$~Ȁ^<Ĵf1y$6l򠫸lIZ}󠫔 yo0Zx;ZueLa~qGƻlYTf!d܆#WaL$D$}!{ʲ~WTND/ȥXQ(E!nAsbxJ' 8:*A#i}U)5"ho0&0̥KV9#(fZ nG-8OV ,(XtNp8JOt)_q%|[7sh~084߭߭Υ޵^XFƥt`aJI@ZAo \HK$BNt)0 l;o};"Uͩ^tw3 8bWXiH3»~Ç$0@KPyA@ y;Io]ŏ3]͍'-fmdqr\cX=^\~]}G׶k5_6`7a(S O"IkH UR))i4ִ lrJQj ;gק#?P|sQ{zi3~~oLǯ\:^e9 sYƮ>ذ8Qy-e JZ %f;Lk&#-|`b{K7:`c*4m_4빳fvjrCӬ`tMx#֬n&Puy)͆xJ u:׾k:s5\^|ܼRu~@93 o{o.]-0D}D!@YXIK] r)LdxN,n SfW Ǘס솰EWB;z=ڟ(MFg32@/-z+Z&dV P![p!0;jc(iSZ~{ }!}sg9YjrۑK-|JޅjJ[2f.!yiJPTSNyrPXBIGtIjޓչt7ځœ8;zQU`+Ϧ^TיsUW }UfE3^Y]Uq=wU=t+Rp97讴+D>wUŕgϮJ Y07鮈47>zRm»_~](#7ocY{dݕ2 mD䔣^#L4e^ZƒL-h#phQ+cLbx_2Z<9Y@77IX41 A'ѱbf|̯M_ľ]]COdV Q?[]j=1XңVWD}s@'kxRyPYg=)㋦Ѐ0ON45wt7]~~ }L\Rx۾)lnuig `g,+dTc,虴W%hM`-ZhE7Qv> d7֘#x6@k)b\'6M&D1+0g!% 4q:tA)MR{+ zQ '7g$J3?)&;b Orfr\gG<+-!g9 l4ִkL kzM`^;ro{wo\|p$o.|fws5W\w7q\]&_/e0~Usn1_~0ݻO2{g(di,&}WS~B'k"[WA85ˌKA(Å5J ?`0'7x!<7 [%"0so+B,>lf2d)0X[r$9s-eE-i q5EU_0n&"7т A4j8Wc)p0J_Z[UX%ZW]ɥ" wV9-(MрޯOL=gF-ډMҧS+i IK3Z̛oAo{\7 XU>ƯQ>~8,^8("Fz:ki|idk$-;G摎H:*H c[E+XVr1ٚI0 ~$F횫5?nG?ᴔ2"xJקS^yP[_|0;_bmzѩ&缳hx'Ď~~۟w?aك7{skuB+0Kԟ";ǝ\?`k M-+7Z^,*Ss >jԋ _OşP22-c[';3*U(+D6V8E-UJzqlU% :x|O8T{nW1[S8de XW'^]_ғ-db`NEfq;'8=,#;ۨ3QypF5$%QrAyfw1$jj{OGiE>4ڥ )+" cs4.yR+ !\J; dϓ/\BU&jnkT &^EPKYO7L1u̙I+({)7ӻ#hB]+D L9-Szh-1YV5fGIWɖ-,NCR,sRId@[bJ=yr6ZűWNx8vCxmJQLZNÌbIo2;e1K̗^B < ʝ=gRfK[gxz}j,·@< Jk2]y_j91%H7!YsHZ#yz'` 㓳}jta~CiXÎԤi4&O!p h;ѧhQSdR?haUxHM M `FfvZ|0mZV wn]~n98uB֔e.Q#)>Mz9]fnh/QHkGH4= *sgAe1[ XJ#6dw դ)ie[R^r1K,Fg29fj)v]jV#5s.ć9>\^SK|kas:k:Tx QtDH,] IZ\\Pd巘P,YE֟uv6̆C)WLŦI9ٞT⓭n ykpf6c=)+;-yO{* 钵;?, jMiQ:Z$:h[nq1(d䶸 qM&ǘgV!m6fTO 6Xeb&Q͹DD5 Jb&%U[3V#v͸DVmu!Tօ{ՅNla;4̎݋?op:{[/f #"-3`I;'R;Z  GѺ Y2`Q&hZF#\4=f]ҕRf9kl? hjܱVv`זC6s#%WIfGP3fo Q\AnLXBӪì,dB\-ApL C*~i&}s>leE#V^#QhD8@*9q=9 30-C) .@*m prU51m$II.k!+NiC KZHh ‡sFrU{ҋf!\tYKՋ^Խ^p59 Py5Mr0mʠ mNCbPn(u䰽^\Ž fVz"E#8rz6v(z:E+wܑ;e1A'~ƓX@\]8.o|Yճybճ蓷8}wi)ѯx<9ٽۘ$ŏtF.\n^7\ڥݨt2EKf@Q>ƭ-R_ Am^uGpAoCU& ]4v/uHVaVz9mQW-L;x{h,_`;h:W/i~ʟNRBj,qp6>z2 07L'G2FnqyX>ZVU^r͇QV9r$.s#sW5Yx7r7X>\Δ2&-Me@hmDxeL$9s%3D+Zs@}\H 6uIP# :,ϛK%8ol6{t79uoPKyqɷOKYIH7OF&ƃFs& =« 8*Uf{.LɌ"zm<QK>#L%!DMqX&4aW5Դ#8ޯ"v[4׫7,tWﺋ?qT|Α#jk0鄔ʗK4-% pH1j ÄX^x=%ĿMk`t_lp'heVC|7;oюݛ}CLp6 L!-:v͔8dvGݕ'͔yaN·-7ܼ gGl W3m /[^›w]/tuEZҦ>h5'XK9AC.HZ7Ժi[s460^mԲ4wnyj|{ zr=?L5-%Wsބ7ʏxI<qt<]uθ? ^S5_6ݴ`,4_hlwB`d#zf9[iJM?fa=r/lwEM ̘&Z#P)v"h Cltv8os?$ E~@ }J(5%s))m $ˢO+B@!8y*qtrkc69"QHP 7F4 Uvj.7' |٣r\g9lSMn,wcYZG̣IVF+Z ~heAXM$#7鵪qe D ޕ$ٿB̗<"/l7= 1b !OcTRn`FIIQdR,JiE2"DcBMR^dɮfA@ clP3I Y\%V'jΔ)rT񺵮:= :"5%q玖?zUgqZv@w>jD|XIZr ;f3팹@U(|HŅ!6Q\\ݟgRelK虡D+ٖJIͩQv]FBLIY :Gދ\eL&M, ]&f:G w&imF2hkS;/έV z;t>;o , @@ZJ3♕*n ުی<$9xLԪEpo.EUlNlE^t~m蝍 VU^w t"GT~4eHWz,e,-|2Kɠ!_aP0rFh>}#u*%Fu@D㢶jV( 2QyHt@,QxtI24(gmC$\(%+EZS 1PKw9W$с>ipTȬӮp`WL `q;TeUݔwS+ψI2J$W}?~'~pEڧ]>8`@yGoφ88s3FG|$xh.%$@+"k%Z3t [,\"U޼JUU<{Q—A",ܩKiuR#Xyp;p.sơ>pj|h@8aYͲ^Юupeu]4O`tv b1ك7=Zst\pq1SۋשSknx{kP١Wk8u7 I9Mܺl?ΦϠSM8rfǭGo1f<{Bse0^ܱeUotRoƕUAfu'q~ʿ4閟U-]>aKtar=lEMܳHSΊڈhҬ𘰙NϘRƓi2~ېHiv'ϝҝ: Q;׶tfkhꟽYl:_(GEka[(qR 7 ۠/V( $OL-އxbR4R6&DAY.ٽYcOw ~/m˱y[iυH!w?l%[͓͆́֌ir8,Ӥ0P4ORnAј'@ј E炈P4R2kR6?>&а}̩R.+' 4% ޠ~;[YJ[S^~.K݂HyReoxq?R\ۃ R0IC2eX]F nT:$^jI|:w|?3\uf,shke]/᳀Y*fEӬhMiV4M# -u}"Nڼ~olÌpօ?dZ0~鍆q lM0s[YBN5% ;ZʏWoyM]YcW: -\w4(H9w X#1-<6NO?OvOv%$!o,]oIpi*Ӵ4bS"6Bjr\38ms>n)5OyQ~lPyPʘ5W+|E"4.i@t#jKK8,ҖզmYʻ4-մSI26WŠNVT>{X̓& a 5)0kd-!8Sl`S RX/NRQgo :z9^q"!J?d011+)Q" $qE"$\8E؋EXOQ C5 6 ,%{`dVԎo3NQ#>H, 6sT+dBYEr0 UN]PID,uT+T;%D4u^ -(jxUd+TʕiyPF8ͽr>82^',ehi%a3%C xä.2QFjPϫ$^+iN%yݤ9k{5Zh+xn4HU"CE%&SmG y ٯڌ*jUJ Ti*-P@տ@'1wґs)0TJ5.A")S<Eq5+Q=GUޣeBNʊTnɡҌ]fҌTJ3RiF*H4#fҌEè\NXA*q,:YZ4"BƉ`|siiiqꏚɄ5"HjހLنDsQrW_Po2̖17BHeuTݜZ۪2acJʊd /XKGh 8׊iFspgVi Ci[p<ݥwڱavfuΗ .%34VqcVexb)cVU,*[QEɒ|ۤ蝭kiw'ZuсrD ,tr9BQ ƙL%Kt%QNJFYGeSxh3 aҠHnM>hѱHtEO>"-< P18p6V;E!2aN P-&Ά|8M(O)Vg7/}cuٯx,hopr /&(p]OJK2Y'ˌpCaTs+9tIw:]ofΦI\ϔL)ѯ\š oTFD%Y;;!SP8q0NJ ]育npC3iQOu"qI hc5D!msZx ?s7\K&$EF:Xd;nc`p-oKۚ(~7kz}ݷjh 3 oL%]np8%$^޺RwG&4#٦9w77uΖ9܎A|2܎DN7ms; BV]I+X~Ua+/Ԭݩ0ᨵQ8yJ>z|{EzW+nrUWݫ܄g/,p}tLߌz^Έ'7t.X5'8'~?|s姏Ç>|~"s'A?o~IsgKf64w= +T[^q%#n h~ 7_ `f5ƶNk&Ftes= 0 j~1i{~zV/J%@ZdZу XnY蟲uy^|1־}lq_4 |`ˀJ%iUfT-pxn&Y%0F*q%%_^ PW]TuAz*v՜#“LWñɠ@DdTě">\ރ}@Q+aaAK78DMrpqV,C-oEjSI1D|^g%SŘКZ=&sRU6F)D;MM.xu%QMnQ1Y)a m\㩱{8*.}HCLI;tI&iƊinY<0D$8V)]OD*Q)eZ*D<rob\`T:RDb<,緛+bUW0֕b W(+6WL>8'۬픟|).oԒgS{E Q*o AAٻ8rW}kXE57܇y OYhd{ =Hdǎ(e7`{ӜVu7v}DN'Oц%B x灬`szoh Ɔ}VJKxu!X}NJSW+E18MAj423g?2f)Syu8 q0₅k7fD\\[}Y勓9`''gOޯ.~pĶx1JQAI T*`Q/2b6@4 PF6t5EaY2Sb*bCr,J[d{9;r.=lvn0j#B1Ԥ':ˋ -oXXJ )HE#GiC9w@JkJ5m>ڋ瘘JsŨ;*1aõ8D6?D\E qe@,j "W֪JXhhr/!\_3+Ora("ưj!4:uI%eO+ H*iKQMytV0s#wUuR\\M^Yr,.`\\5㝮`Jκf-lR*O/ILu`C-Ź1< :\s.5 >딹 [!'><9`ɿ<95$.Dhr žǻQ2eCZF.R^Rh"[Aqm 1 |haJPaT+_,:iIw:95U&[!fo9.e؋P5X4Uas߲d,b)z;ň3%MJI=^$1 f*r`[m܎;2g49l|<*plРCU)c6MJ>%V)_/\"}8&c3+0h2;g k'ۣuȴ2] =D]$,qъ2_58}E\MDK=ԽRW C}x_%y |o߯.ο^?'|* ; UN)qHID&YH]E>ͨëEWceDlY"E6WnJ ST66Pj +PY)9Z5OP!p9d5t,C*ɕ^N驩Μe~*91#_\ ݸ;/n޽Uo}];[ےLSwoO9/} 6lq,|%cvپ4nKIc{Z F4֟09?ri2-b '"+nZHhSqQՋjvH09X9{(1BN 6w5ű]I3IYϦ\/B%}9f>ja}3rię>Įv7UпuI tr<]9;:rґWu3?ݙ X97Fgg;wut׷k>dp{nN%>.#^s;ٮ_ =4l'-Wٲ)>Kr[i+1c~|]߂|AdhwAkcF2gS|)~xw9Jfn jeg[K>BF⥷ձ`ޔT{(X9  \!^jmUDw bˌQe3:]m9=^qm[_rܥRrX(KBrwrvv?_fp]"/o~AoO'_(E;4(Fʿ/ )Y4 N=,(IK٦ $V-Y)cjjDBjP`= :[/(ts{)?j{Oa ~Q\XNWI4=aD+%v#6|;W&M:5 ep8\RifJcFRM5a{N{V*ZD`65lSVg%V]̪Vn!TѮwWݡkS<\|z;:%&3.ѪZO>6[ weraz[~Xk0U6Q6aJ.Yכ$/,tP;de#\zŻWO f:y['mvu+8t9Li_wrNႅ(B!U[S ٰx a^@t5\㤬Q>W؜WW:/uM(u]Wn[owmogO }f=[x'^sT/I<]?遫Gm84k~g76K/(,bO6(K3n%%' #Ws}CB}3{{Y{=4)4Ȟ~@F޶O/Czf8,֑u'b $[-2ub#!YL;Q+驭 ZUmBAh3JID,:+!:̜=_ ͍+Kֽ~@CQyF})w,(^2NƝm^B.Ęwf`&BV M0+{iL@dKC =5 RZ(J- Ql Rq9qWfƱX,|T,Ucg[\ꤸu?aX\ ..`Jκf-lR*O/ILu`.8<6;悇r> l"^P1v- wãWKw2]>4„ɸ<)P'%xxj ΦۑlE9%@')T&eB51IdG߀S(8!1 TrJ#I2{"ptLXC>cb+IEeGo3@'^QIw7u}yty5>t7|㫽AmHh^/?Q_Ot8MHX]FfԛTU\2eҵ\=x0ϝ՜ϫσ?{]sv'EEwem8UxB퇘7N[NNN~@YR,ZP]xh U?v9c4\]5L}Ѹ\]NCuc{0r5{"M>jrFuhVK=>a0r\xb6gԻm*oCfu`EfWt/tH1?8HvNz=}bV/}WC>Bi5E)5<ޚ @&֚ p.d'MvJxϻLcX 9舄!bsVr"n9Ȕ̦b&mՇ_&l8'^654lAxc"kOPlB% :lr&\ldL_ш&iھW{cɃ>a,ZsoôO=m%+V\ӽ5fgugς&k{it uqcgV i͵(ګ?&p5YX'V}S%:LWM?bWwخtF],.kyuf5tbn C񕣶;*`6!zC@+r& Kc*H%e\ t-ζeXsj/Kz c"C u5B9!#xYRl"f :`VvY`W0rt5'Jwbh|rmAį=XѶ~{rQ'nc3#1jeRH\P{yW+9os j s_þSlNC%8GCsB#eZ9P[E,Bղ];"zrܽmS꽋ѵMlWL9"ap-r;{̤57rLFXCR9lV|vBod=NO}P4DTADtĊ%ackbM;ϹUw D%6r-%|Ml"e$CAJ^w^ "g7 4Ȭ6_\[B3՝dž߮OSg: 0>KBC:X9k)ewyz˗ãT2RTW哻BU FUMaQ%JQ5>aBo[¯u\%H R.L޹]6łbEbhMLfK#c𶕩G˜ rhЏq3iox+_북Z{\2,nӖ {:$; >Ӧ.iuM.6NVTxFius 3nvJ{գx0t imJX9O5R AGm}HSt ٘fja̮>\s&;LurDqlb\ t1}^EM,[ d7zZh8#g_thrBۆªo_㽨=M+mHşUt1Bo4c$Ҝ{h*ƠMkŋ j8B(Ym00Ix|#鞋C6zdcR7;2k(va vrǜc وoVB_:} on㈮D!F uR(krY'UsV?UfLբ $]jmc(=o1E1`b8e4 hz !rKzj&e![:]~dɋXœr꽈iJ96T &x(+k`R*rswLg4[.`XNbIZ_ήpf'{M#;Hyg,|G:#ׄ/_/vx,g)POZ-a8g715lϵjM8qNY\^\sso@(~+>9Pޡ56MSޛj<"@ͱ&.I QzsRjD/o7jKb;ɢe^jkv0.<{]uÃJD\[>WUTl Tt(֢ڬ| ZrBX1]TtN͎1\(ϳTWut-Vj? 埔%.. Fq!%o,EH@وgkR-6bPF˚9#]ML0H5%mXB.iw)M ZϾ)POM aSB~|Ϝ?m^v^#C.U2hgRu6l|8qn7!$*Z6{FQ MNZD9S&hVL|i%! kS2ڦB`*5xS&=b]^Xh{9 H]Y=Yt[1cazcoc~:j} WEidH~k\M/g^B/#h|DYNQ߅߅GU3 Iׄ0|,7.VlW*sPKbTfn3?:(gv*G ΃jW[Z-Gr}Ws;,ㄎNPT%E,Fpb"hB.z})4ֹ )#h֑Qُպ~23%XRThQHFԪW|>m<*d9!"t؈>{OD10V9%sMg}IMM`rӚgڥl)yoV|8)Čd-U'6/Eoi%ꂃa:\ ~\g%G]Lˎ!۫_Yp?][/K7<-^+Usy rjN#WήqCL؛!XaTu $7_g}N3,u%w'2~n-D,/QY:K]?w+%{,=#+ld~? 퍃KrA~YCG3E$e[^~Ç(d%6lNTU=ęi>cF[o$i.eq?>^I߳?{7.3+Ɩ_ng$_G'm9Ԡ9QDaퟑc8<Nb2E5x-~E psU^@jm.ɍ8!p4e>4Oz_G.B>wp;s*ggWS FZ% [3\.~>J|1?ǵ]Bo~ۻD{sTۺyݞx~>=x cS?LlfG2i[r S; dHJlI#nmFPqUVyba8ju`żDO.!7Y .G$,$$,}1| jO{Yko y⟳ӽ?;Av_9zWGG_9B,{/޼~pvHbg܉fw ?CsSjho147C ]~q.]#>r[+k+@v0OBvHm9F~fUϠ+aWlYl?-Zq*)ԥFU!v0`_U_}I*z%V*pC}u̖T3J'H~;M2\$J,ІnNѦIx۩q I H \ЖY?v_G#m(,q1` NPEǥ'\xK^jQޑ~nZQ%ɢS~g!ZÍ1%!ʥ-xc !`T׊+wS'~֡'_Ks%4W1`M:(b[]۪^4 F }>@1XoB#CFpxl ,{b 3vڮ_vEE}BH&O9(k {*B mKJFD-mKZ'`C>rWbg*;TW|0jSX zgdK<]'s b*z(Hb#T#*b?sŘ])Um1@?~}.pz0d}`>9suv@O` ! K%Hΐȩ"-dYC̳V,WfYs1 9g9 3e!wPUC=~_3" R!Db:Q"u;?6 sJ[Uv6;!o,bC|:{ DY"ydR0W$H*) Y'#y!ɓrm=C)"TxkPlgI@rAxYPf/kbBaA-F=ǹ}3FX u㋎ VaKK0p:h;,lpǣj6{峸#ONxrᑿN9~0,X{Æ/7?MjD+éh4 Q^erKrf6i "Rs_; ZZ64N1& U1EѤ<;Jp TiX5c9RLmu!/ yՅ ggd!qvݎbɟ]и`qПLqr1 DhIAYɉܯn"+x PG$J$p"#{YV\頄W"jfNhcb$/H$hSXc#nGøbbvEkm,e޺O0jᜠ$"5ry$)Bt@kxTW\u.Ar!L+To$h!#2uMfB,q>&͌J'#n}1bF,U#V$4"(QHH!QǑQUJ&,M($SA!NF@&)h'g|P(&i*tbܭ?jzq̅Kb\^ԅz[p$L&YTħ\MAd" *p$^]xCz3Չ+pPau7Ty_Lpڮ|$˔46Ip5+e?JCy:+/`{_ .7@}c2ϙk%L'2$B&[e3;)Jf#Z""ě@iF): !("$q !JBt$-- АHEE`Zs(BrcL.b,9A]`c1rǣ.m=KXz|,чaj9GV>6䚰w}8h԰D9'2!φZChJQ 2*d`9zdmӴEV/&*Bם]WP57_ɾQҾ_(O Zf#G3FΟF|kI[&_kiȗG^ɛSCh| P>B~G΢>n2//wVvQ+<)EK")ˣ<ԅrZ{fx3A<[||<8|Nlؚkj}%Kp!CV)6] SGeICr@S#gKjw}yCvD BSFjv@;n|d 58K8JFEb&W֥oc6RBT b\Fg}N1*WV"<%OSJs7.!-㧧g $^+ܚKk2o<ٖNl/cq6sK[.2#wizx{m\AZndC^:+v`x=l?kPak.hڷLGݝ'm#0$;5mG怀>6*zY{kfU3f=GDҔ4B$>O aw< .V$'y֪d=upR,ׅQ.R:J|1d)kLHBe82,h5jk5{5qzbN'߾|r_NB<v3X^nfU7b],lg \䴳b ٣1 4[U`o@4 4ƂFxm6f}cKnR%Ok3Ug?g=.=uC}xv5 MyL1ݸb@GB<,x6M,Տo$褨j!$BMAL$ P PdA@Bi%㥈|VB918L4dX3e};T-b( dz"E,HL=5 gHbP3@s%5l(иUGht}w52nt?scMr'CXd>dܙ!fN &4 Rpc)t.E \=2b}]~GKK79T\LFWq<z hyۗ?G~6OE{.}k=y֟dޟf-Z]c~hS FMV>qo[!fyNV!'[z η>vږz2=25)խ#?9vyPBȳLYBm\\opwg~yw;ʷ;,:;ekI6#1H (5$ qf`uԓx.;dŨ\!.2 H[Rb(-@RZ˿UD{9CxL`ՠUpܞJj{bJUYݫV4}orE?[.{]G]S^ڡ/qy*.ݿ<1VkHqQ[dޫ;0YțQyhtH-U| 4@i"&TF!H3vRRg-StAj" r*ygLT&V|϶*8eT[;:wPuFRE`&XϦKY pREJ&`w,eʛ$708~sfOdGc>$JDH750 Jdϑ"#i F $^kўahz7+5KqҸFHU9E <qQ]3ߜ\T16K)nxnFO\'p1A"q^#aYͲrinS~õПp:\{ ?93MT>1`R`;: )RXΠLq'A 1${ ܎n# vkϫ%.mtp-..6ʉ6=?sv8fCF %?dP411+)QM\Q(%=""n0  [Sud `΁V( 0qC@ 5&8Y99J}t,YkTG ,;! ?s*^,{GtבWlGnvFD4A518^W!8qȳ69KC i*ɇTg?Dip^ -PNcx&ӻ0N3Xfy:ͽr>8J}׼ӓ8MtvLox-@$SyJ)atefF6[7!>W^٭ oێϓ;$wߡiqpYAڽ3Ĩm1l^~N;+:7^ƶڹ31 eR~K#LnJ*҈L֭!zd@+&MDc>hNu]vumҭ0,9ž[j?+ڟ֢0:{^[q^=n*ȗl2oY,B˖WT^m]G چ{-C1AկTWYn)ocd6ѳ;oe/lxީb !jƮ5ڙ[ t}yS=t[*f$S:r1"h< N ك Ku"/^OTy=1;pi*)U)+^Y- Uv԰zѕV)X6$-6Z:ԃѫb钜ueuj~+w&:Z{k@d?M+҈7҈ j|ɏXm),NM EIOE,pNl>!8U!x?gyUkZSZ.>y9RvO[a^cB,h9ggwYw5q09^xU6R0)yHaEnA 285̷` %saa08:} e*u.ܗIO ^ Ko5]ߔuY̎|{8Ɣ1Of7rl~4ǝ޿jysne@ڡ{>0Ejd~/Zh sl_-=[} Oz1}@_2ޑoWMd|Myز"E{D3$1/[8׭9+VF>#'QD FcC"ܦzʢaeu ܧ'NFTF3ؠ4el vsEZqH!f"0gcҜ*jydR1why*k<sU5gs@O\)]GsZj\eQ,*>UHͩ"%\Cs%1R!\q=sU5'l0W\IN oYO'dRQFc/KD/tW$24f gѧWO | [ےӇQ/MnlͦPԯ iM_4zKX}r&=g'IKP K4D:&jH9;|IԮ<(e)k'Yr2L*{Lਲ਼Y)sYTr01ʪ &Yf V;oY?gL̳!8iFR2R)a1_gTy2$Ae71bD q挱:'rSi"ePoX`7^3`0@BB,1oS1.0N18llubhF%)MnpR9&+mw$gvrvsb:HWp ϸ!mɆY7}app[ iQA'TZsѾ|Nӗ4% >pmb_̵dgIDѹ.'QJQՐhQEO"KA+Sp02h;(SJdV仙1̩!q~͋wxcsUsi9}xݿ'}#'0e4ᙘ,RF̝AXY" "va/; ' "H 8)m`.z\0Z%'/ώ'W+JuaYe C $Iqs=JOdz-ʑNp !Evuo^Q=q͞ 1(g>RKb;Gg1h$ÜB"ʹ<,uQts<ȸ`@LÂф"ʁQ#s$-ɲt bڲN"y \qPtſ?wKg6\ҚU3VrAtB;P8#q !jU츾51wLТgm['5~>S/%Ԇ/ <ta3lx V\2#I# £h!L'@c6 ƛ6'o$Jh5'&Me#{p\2!-K[2:=r8, H[#Kƚ=MWFA)MW<xr}Kl >\iaG\ YiZޙ4;uzϏ~&\bt{{l4͡P[aٖQJ-/]7[ vo{obqUmwcPɖwl=4۽YmѠWƆmnK7MC: /+-V  ZirtqQ!hMi7iXs3 asO%Imzwn h6n-oT@6D=j #@[5n]cmMSu5_NHu_|w͋sݫ,hv㞗4k< ܰA[5aޢb0'mǼLdֺeL;0n)emX :ȮVNuM7}yB9{=vwmqr6 s#ZD*rRN`H&2\6A%3'bZP``v:@۩0O8y}<#jnsqTM[>d,*pĔ+/܅Z }%KЎ#اUҬz*؍4:(9M9JŨYfց(PLHiKCL*!A+r[q!]v%CD֛g@-ܺ}])e(*/79A?7+̒RI3cV|\N>.bwզk1.ۧ3:~q׊WRa7 fxR6d e+Ѹ%l5 /z1U-^$U(H?=~؂}?ORR*>lFwsVio5KzeIZY80>P__x9U^u>(HB1#][ǡ[{]t׹ԍ'[ˆyY>JWݛrRמO0Qimx8 ߗ4A?28=Q/ѽ+&Gػ]K-MUI.{#_gWo鼒zV? KX)uzFI*yk+WAj9{ :bF9WtV(U 'ݛ^̌\o.f ,JsgӳJw EDhß.g~_cM3ih Λw&7Ć ZFN+_4=޿p M㬂muM6= bI1C꒗D>cjB+EhXb jN{qF~|so|ɿ~o{sBX^:uB;p%px~}{L?~jiۚO--l156z[̫L0Q7[]+8Ykm匿?oXJ%]FkO]xqfV/aWzӊșgCԫS-B,!`@kUg94VlKUVT5ʹz" 9}j7 ?ir}!έWhdb)2Dr8u~'g?Fo_Rg phk/PDK04x O^ 6>zDvJc[GY!YH5<}F;j`A1}NaYGmP^+8 ՊgZ>+Ҭ<"fvVOox+be%51}4b 肍+RWVTE2aC2l9)}(v\KV\6YpJK`mNs҅t06&Jto*Ż~qt߮X}}ZRj- ||V &DL4AX$fUQ[Ug$Hu˨!L:l~S53Pg_PVYwϳ8f]Hw5 O|*L~PܚfO j6]9LʁUƄѠSb~ RXA(Y@M 6E2A L1|;!:qh28m xb] e,,fBL @.u&g29qw; Bu[[Ζ F}HFh]>=tdd1%P TA 1*JSmفc43WoU %k"Қ$Nj2Ћ#B!* a6]ϨXĮKI% _>ektZs]5L~}[2M^ef]wyoGgj{ɍ_){C 6tXV|fC[-lKrKF-ŮWESbUS]rkT-/帮(*˅^ALF:v ?exLXy-Pb+*cq*#ZXtH (BOkBBS 6* YmB*yqrOKiơX c!X(m)3-'/ٗ6g_fPy *;MF#6U&.yd\̮v18Ea=j vg2(n-$"jy JBG)XHWu{P*2bHӊaRY22CE{F5!/iK$1_z$օ$AsQY[2NMiƮ b1Gċ@D"cc|PѼeJ]XHHsf :&!`UInFG6V80VXG/;HHQ;$Y(@EQM2̤4 ͉N:Ftua8-{q)v(*~ݕ)s*H`9ژHy"V')+A"v2Dh 4]s7@]%8Ј Ɉ0:#)qymRJjcHV1O'9zSNI.6ZTVb:Jn 6̈́ S%H4G$Ý$x$`M$xD|<ʝl9{Nޅ ݙ+TP㓅or<;x9(ãjVZzp~;4Fg"~mi2L}\!ԁq_IZ5'p Sa.*N/ >ipB-Tld-~DMTh;tq~֚ÿ>B.<`f;B5Kz7<5Nrۧ_?wx ʲo[]emc#~lفDGi+ү5bX-DjF[mϴJ2-ڒiw0o!'VgƜ^ 0繅 u q ]z R B UK,-],=\A@ Wr1py)puR]Eb $B9WY5&qU \q\C ~)pպpR" •TuApA/E.@.PZOpR/=\T$vSq9`W_32<\e)fpJ0PWK?euh-5YVK*t&$̮UuYM4 ~+@KZ CeUN[ IQ^;|zK,9vZɯP7~qZ|0O߽_ Y[Ձ6#]pʀL N>ǕOH,yFWJxC (K[Mm8"aKQ|4=L'ImdQ~`jm(l`v7{v(&P_Am?XהckV@f&D4?|\w?_bUOQ#%UdK!YZʻNkU=tZ.={G4g{ 93\'.U煫Ug{q+Kjm?4jw>qpV7k[WiiD-hq-8>㒋M*Xq^5)Of0j\i2sy5ޏl\Z]̥i7ױ zhUиV>HkaS=1T\O:1ګh$mp*TT,5?ES&8g)Q(CcSm:+D߾6SI ly9_)q+P\$Be25ZN,DzMA[ne iG]*PRI?|e+ɖ`r2.-m$#)d%ifupfApep[Jf$pu>/pRm]cWv gv}UńjۤmKMV~eZVj9[8-frP%;W%i QW^kjE4[qg(Y[-&/p''GI*"pGUTqf sQ2*;irz,o)1ǔ0!^ĨLHqiE5Lڊ#M}~3N]#K b:6]&asWبFZ ^$,ܥD PT+UVİ(WVX;UN\~! oa16}#NW.T'K~P WIb/'_L9L Mڜ$u>sVVWRy O֏[l-(A2zIHQin^py尿R 4z~Q&Lza=<<;yO7q;2n;]kz8N8]':ӈ̬r1vF`0Oq\0Іf&yvud4mϐgФ8r6/bҽjD2X Wa[\կopkpr0MYTzyu[J[&?d4_ۈm]ǴoO-61YuZi5mnV}a[aL#_\|u-Uj?{[Mݩm|Qyo;p;x͡%l/'NcŬ:>@d(o|7'O>~5O7o& n4\ .?a{ HCgkvl||&Ȏj}֜;.*ܯ/mM..RrC|nm1fݭmT$doMT={|Yݨn$/Lo[?[*l*Cs'B|fG槅WR iz+{.b`nv3]hO\yC9~)R}WiSٗ<~ 7Դ>!:ww~{bu!QsX=wl66R |L4;!ΊHT@19x@IzdǓX|L!Y= ,P b OZEZ$q?{OǍ_iSx0'6qeǣ$r-b>V%l+|)}a JV1#J9LD70&Lb`*VU)mU)9|*4 Wmm ]2j͋OTig3jv/E .BkԶ7PZtz$]9Z NhX EY`3iO%BHHV)2N98>@k`]/GԘ69ilw>eǕɿV>$yQ<=EU=WO+Odf ФÖ>{i ꅙ=zq])0_ m]4\mtɜ-fG#՛.d-Lwbg i'Ltg;n1錠[% I7Q$֝xAޢ^#p2QMC;ޅ,VWϻ׼|;;zwM'}uSLNM2*z+wx~z5BP8'Q:|w_xo|Wf_f.޾_xfjku H7 w\ySm֋üڶ|5^C>|{0$kO@~<${8M5lVy+VaU*Q^~d4 NGi\J9(,ı,R˛]ӏpV;C+3\J,W7@ vuGvtA#,r^gk'az3Sm٬c$(jCz&hgt暄|9,XJ2(0^TQPSQd| l2=9ֶfj0([itI:G[b(n]f<. .CZwdsA dH4LUDYll0*P ɈL*c*b`RRE;8"댉"xhLg#m_ĉ) ZBmIVQ[G"vV|EDYhɥ4AIU1@-I6῿/S'jb([YjF% 1epNKK)cZӘQw $@R ,,. bt1>7t Ks ю9s>,PfR[[eۣQlɹ(,\. J.9msaRˁ+yd# )5L5gl휱Ҍ6ÌMy!40+Nf&q>[Pwz6Mtox|a|8}[հ``+DdWa, B5)yU& F*maN8)y%Qɺ8sx<ɟµaǦ\[7zڃ{c=tEZ &VMЖ< 6+ +1D$)HVӚâ ):--d %X5b^I$HdXqDfΑ s;?e#6Ï#YpDM$@FM%۔Bl`]'Q4Q2'iߔ#z'06& }1VBf;T>sLɐl@ؒVh!XIԺz3pn:2_v &_g3,ٔ/| |q7柱"SJU6"b +Ɇ6؄X{$Gݍ+|q&;vl ?0XX}35َӳ錎zy?bޅ`QǟLs'ќU27f{#rQwwg8:;>=Yz{'/o!ay2G-k4݇?a4ˢKTgƺ+6RCٵ-.E..,"Eᘨk,aXHcW*LAِ \ZKC(uz_ DQz\**F5V@iݶ8̡ zdڇk,~"'2Uc"%Vӏe 2--Q{tD, E]kNcGKI.転t Qђ/UzڼsȠ}蔃8a2d%AيDW Cϖ~m]Ÿ/_e҅rl&"30rBdK*5T(ڗuV'0ULSej΂́X¦!P7RkMΓ: Y@`L[2bMt:ݒS}$ܙG]y'J?.5W]^=A\02nuu3i{#޶YգvH-j9e^UƏwޱ0L,ozu/yJ?-ϮG:8d%gl/C}[/9}i&4̷)q :/Bzn7!܂a$H*p$iq9JT7ȑV ,*";uU{TuՕF(}@V"qHZꪄgw몺z3J)rK /xu;%/ͻ.wy@nxsB BpId0*6 }wQF\0N܆_/CYS`w;]$(TyҺ4r W?pϜ?PkO6Ü>G41nۭm`H<.Bgeq"'G>UNM)r@#W(ws/un d)3In)Ri-wKIRZVߢ4LRWE`s8qW2<uUvH)lUWoP]Yg"MRWE\y0H"}PWuU]mFUUWVʪ^Y+Kӥ4-ЩWVʪ^Y+zeUUWfw@eU^MeUUWVʪ6դx6;2W*>+LtJi1M$!CP #3OHm lV_+|傯\fr'&4-g*|V._+|傯\ rW._o{\б"H:v_K@O& xB0CUݿc\FYINh;ux/y!Ԩ 7AlPx/BX5DYC 5"'pBFq2{%rrNgyID̘Nu^K2`1~Q)& hS\:T`ƹlϡY38.4dj|}%g^8~x6l={GBlɤLax8j6$w#9UJRia'Υ!8t<{]r bU+kwI3#ii)ȠEs=y|/Y`hc^vVr^(BY>xIMkxN|CdޣFȅl!)Sj .Qs_=|=ͧ(yF)<-jtH=f'Rl!(@%^r 4t3_Afu Z;nJ0wQ;42G/1vyq/Ӑ6ݸg3GT$x Ga.s}R¬` 2gd|[Df$d,݄df0嬝̤Dd-pfI(RN뛫gڋw2'"}&H_BiMŸַ5tӥ@>/i< ׹KKΘ'PCAt&Ý%4uϒG&0>FUdLsefԈ%[bb$WwNy`裱&LomTDlTs8o|H7q O?GamܞOUvHMFK!SHN-e~9݁Έ)p,1"Z YA m"gnueь Y P5V4B4*P s X3g68knLR[Ge#KJ_v^IiJ? ɇ5R0ǥŽZU,U=U6# CZ4 ,HZe̒Ɩ3,^ S e߫qšW9'^:S daJ *j`*I![ V]`.%uZ,Y>CC $PԷgwpzf}!{K}(񲀻F ]%W-WY*Y%93,y`$N#" HMS<h?0|xApoë3JPv19.u2 HmKcpms{(ei4?4Odl/+ټټ wfcYw$)~4i'~$?֪vMj Tx2MKw/1A-йi{'NEEjN -*Q)3j[]V'ӤYsu.5:iţG :YI N&s ޳ za{ (s팴L+TSe1 "Ӛ'cŘhEqWv͍(揙=nΞqf:(I4 .y>)S?}7Ӄ]!?};@(&l $DžhY4R=cy5vuxR;IøaN0Zu.7Κ,\HXN] 7~K`.eHגs3JaOJ5뚮i])&H6!Ч@]gַ-}aWE}K=Y ڢ5N ] SO vwapZ /jzup^\ߌ.n0闿&7#צ;j{.fft+*OGβ ?Lr쩬[B\ r5>'d=^);r %;xzYJr jJs(MYITf@'$Q;L%ʔBZs`L;B&H+Q[e{M5C=;*Lk[.M/OW,|i:wXӄ%eGwC˵N%$|V*V,0.T*4|GA\/p~3&$|֥NyJ96ۊl,`\ӊHޓmW|i 乏"uMM[K agl!2Zr3_y+ IqtH?)a'Xzl jh]pbݨSat٤sZO5Dy!0H^F8"Ybzo_=? .cMJ3  &:b&GGuF^@ 0!mvV^I;cRnxvyuLJcsOaMm!~ʟ'qQ!s9&}"G0o$8:K>WeawϳfV~=tYi] )0^+}\g$ U$Sxpbz<}-z36nz^u$^{(zا>Mz(JX+>ռυhp:q5 Q+D?t@1Xo@@'0Jg^ɛ&̶LL$6(7&y "2%snPJUHǘKm0!uWg=.N&AE$ZPTM$VIaT& [ɌM%Mh#p}RCZk|Yhu_1E-劒{N,w-Dkn-ۮ)Ĭ.BuAF(,Dnཅ ڻc_t>kuA\Ŷ%,_E=JPLtم3 ƨPB:jċ5Z eQ6Z A!D*y{`%2j 4W:Q Cw)f|Em]?oDeG=XAh$R )4?Li^'z-F}~ lArLi\/րIItIg hẎ$5LYBRV4aI0&m W!$ &m0gC:۽( HV9m"\*"!7ɗxxՑHb),J= $w 7O]xc Tj_$'*T*M&!orh9iH3ɣyA^ uÐmC5:AgY7[ Hy#d\O4DNEY&$&Olbxı;qa=-E:CD6 pȀsY5.ANCt)&+$nɷy?[OHOmx޼O:զs}lz1"|4X 5A omTDɨh:sҥm]Q(׷Bܡ/cX/',q1=J4Mn ~Pz~WOjWw@>]dJk29+J'Be}u\ݡ:Tu(^pa*\j%tgaMIu^k ZTĬҖU6;D8%3x\ zAT,NH]rڟiٰNKlț*zw|DY"ydR0W$H*) Y'#7Fd%G[~E ޚT!$a@V7q朠^Ą„Z Gy R8XwvcaBxUђa>> P.xd^_W>;伙ǓKwl7.UE&#p* (29%9hɹLsL˳K,-UFHB)C#uLHjA4)2R)\B&U3~XNW) aƶ慼煏 Wgg͸#=[?;dֽ7LOƳαYN!FTQ!koGP2Dc7Zʕ`<G#irlnx s=,MtP+5Q'툱h1\$)̱s?:^21wkÎm,̵eϵ{ {$sJ@$#No^ģBu;EZ$iEaJ-d @+:j& 3Ay q>&͌J'~~H1 G,=G9gT@9BB :XR4aiD2rH\HO zE95 6)MyD;1e>FEр4!%̈́NS~U{䋳vY K勺0_T=_g()$x)hR.Ȥv:Q[Łī w/Lub)~e #|'dÎjl'ܨNzq6Iz ~|GNOُW'R Fmv !) Ճ$'+AotZ GmtzGg=q3>A[j?IJGuQ% &$J%B!hHMbpϞ+2{c4R_m-m zڲ(ITʩ"(A\Id$pD5\L~i ֜A3/鼜k0 n]4;D{.O}.=?g6-`ޅwLS 3lnz8xR`lxFDVpĐ4{0ZaS2<Ҋsfp'&v8i'R@!!Ms[ڬ!0j( >p;?M]Xn\Ta%]}kq=7?9ݻ|h@V ȿa"\py&DK+s9lVDf?wݯ7͉Ł-.ֻ|>4;zg/9MW[; k%%:WNaj;*Hiãp.fI>'ˍwCߣCZM 5.TvMu{ cDo~`z՛?fu7/Q|;0qt@ŃAA0n,/Mq܈ S/6|u.B>rՍA;+/j< 9{bkm?9贖hҕud~X9h:TR5G'a!5Ԁ.x6U]eej0M$榙.#KeЄnn+0mvjG2(eu:{?)}F{̂ sq'瀗ިޑ%q4lD,$9Q6:F$Dϓ`0 !$2_+J xd*کp 鋿M({s-:`o4VhkWV#H6!kٯg1;PZ?YN/yg<nŃ6Qyvӻ_rW pàq]&< JK90Gh SfC@Y#J{MbγmԳ,rx@T}=(tqռk[qnj[4ܺ+j oXps=~n5~шř_<[yY=ks\ǼSi|t̫&^7Vi{ @klqv.? Nٻ8r$W?OɈ@cffFc$me]%n/o0JUGKіJVfd&/8䟣Z/jNi2Wǿwu9$σZBgƑUNW'd ɴ*hdbfl&#.ڦ(.ur(c:@0 BgG_ڧ\bEP^EN ũy,geY nH16_r7~pe߮;+[2%/No;R'ߞ.{ұ͢jJ |e5Aj*Z;>Ԉl)#`I ˸v2/=R|Ʊ=ZPIZ1zcF% św6G]n3ǗY=%XN3:kA?/axθ !æ+D E$&lbA{T1/n>[9_s(vCU"bC- 1@ ;m:(|SuHߺOq-yJտ҃]bDfE)BQZ 7N!d|L.it'n!.*t;trUmE~|GOoї\5cE,B.#K!((߽mԷջg_NЮ&]mvOm&6#r7r?9YtrI.5bzo[vUۍ̴i9twgZ5ohUӃeyY " L$Y֏jPÄP$P+P&.(q EjJɹPBĂurqI}W>ɕJɉ4,"rjYǚb(s`NȞR6&ߨnѬlP3S}kPL d-\(B߮ZHN 5pP~7K\UrF9v^vsgE&Zealw}{v=;xaxf0,[߶ B`BfbdڷWpXZ%j)[ֻsMIy\Q@Oa޸\Er,<ֆbꍌُJ7,63 qG?7f%qupSwǒOjONξ/_9bC Q)(C|uR.6&G5X`YYE2豒< Sh[J]N sh)yKU|g&~Ng96;ڦ3jg%(H&U'<R}KWVJ s-)(4+Vc]&j/ BVciTLh}qwp_7qo 5` "v3"Έ"QFdJ#ʬqauEe[Y A2IVx ]1xu$q1:T4fkÙ{*X-gOat M}WZm]g\3.θ3L,VD٪T[ҋe`\tL Eaێѝ+θd'S{<]6*ؒ|mAyғMWacAps3E?z91\,:9f)L6 Rqs|0{mqaʔӊCe`ѩvhe\*p0aSg̶wlumfzm?u?~%Nіgiç؈S~˟D õq;Odd oݧys11| e #7{vW|=3_{>,vostw=2 9qF\vw7 =–$<n /(U6㛓R75v r`oo}F=Ze"@d[0'Z=ddf}n9ߝ&$|wGSZbt52G֩*erD"-Q2Yϭ2_{L&b-=찎fzB7isPþG(Y#'oZO1r1{B-X\ʩhml`nZgU9h3 l Z(pQ-([\E*ٯwW}ۊ]{5{<ǻ}+ݯ=,ݳ̧˷ i5c4 bɪ\u!TRDAN:xBJNNaݣ8 p;Fv,*Ż~>i=)3QLG+`-Fr^hJoQe2ax?eM- x ESUIa`Z:$WB_U1'֋c^B X,dpY͘bS$^$ZŅ=Ú}JX?^Afmҍ<^Щ\OmmpTfX=C:ase&s@"7P.kgC›/ooVC~y|Uű|⏇ __D?>*U~^z?anäDw垔sыQ2P$kW+v8dΛW_H>Y|hߋ$F1<9vdǮlC|,ِ9veb,6~>u_Fz:#EmYsyq\rElwtVbd]W?lgJ[Ey+ \͸!{V\W@Q!3YVռ=W̫ql N*RT!XD)pKʮMhBSbHv$Ky_\HЍL}ؿ|?V!}wxlHՁTFwfspDە`qb.RD. 5LQ[A Е*gM֙*xw,!kEt)kÁ/ hr )S$+o[w.dw9^짂k\~!?mеY/@UZ,m-k!O޿;{Z'aax} cq6\_0aƃa_ˠuÌw`8ߡѫ"WEw|9npu:K(ŮsㆾcRU60L-wQۏU:(EWtITަhe4ib4}߫kR 3Hڸ$i-vܩ]]J6k7YlmRoۤ\$FFb&eظ6$ޭrìdOjD]aśY^3k~Nz)Q[ *?mK{\=/%v)(b0G<1b 7cӉ9ױdŜ\|X^sfΥ1Tzkޮ3.] \oqvGJsA@FCX`4*LNL}8ь"$H8]8\O{ب:ϗo]շTiڃS ^];sg,jn|fk&hJTn5̈́i1M >ռMPe,YT2me\L^7[09iH &41#aG, (Dxq}ej^8ŋrƓZTwv6Xpg ًA*`l@~3]B/t?= BR"iI.v !:矟H4%6$PWZ`V$SI||ˆ}Y|r[b ׷wHz媭= >m [dX.\f7W$.{L,c05%ayTQ͍QmQ'(Wkg=wd01z}AB0`!q ~^] nuk{& ђ9KC)X[U:~>k_`ُ}HK-iY=sQŶ5Hw,ZQSΝ㎓Hf- JE#juTpB2Ym:y73X YNd@-ݴ{]hhl$AnFR pF @):P¤!b#["{Ǖ'q9fmݞ.{Q4[wY-z!Ѕ ^-QreKum\8F/0Km*CA ցd6׾b_be ƀѣ(9HysQ b-ll%N{ɞTqX zT^XXdx!,s_֣ckuDTbf5vyz]J&$ L >p^qAzyzT-G{3:Dtp5D y tvt2 _~x♹ar[\T Dꌺ a%gz@N[FH(򥖜N)hOfx2aOy)4Fze;Ųe!=x|jޠŔ>T϶6毗.q||~P>ec/Y7}E4LreW_?75DK|'8e>1sj^G2\J[2PP(? c&r<r`R[DB(?EB}= ;`KX|`ݹsκs+rHm*s?%UCw;˼?FydHO;GJQ&gur7*8In.e*wxuҨu~O@s=6|_k1o&+oKeHqmz#@8@4)K)&1(-."ӫ30})4QQ>Ghtӣ-&s{µLelw.waK1*m \Z$6@¡~5y<&#ƗQqt4""4eQ 1dZ a LMg8FҎY ,wZb8z<k g|w%]W 3_;pք_L/O+>juV!ǚk^Bfɍ8Q536 Ұ'ߝJ͕`!LE="@% $MFqY<Ŕu;"E Rr*BafXTqXDI$@ *C^~q>Taݨ076a#aT`X8,Ⱘ]km ¥znE/4mTsG'A+Hyi@"b0M ``Q10ӆ!&(Im@H'繛?*o̟\IvʪÅt{!Gsi0D.¸T~0Sylǥɫ;]W*7v K)NWGX,ɢPc}2D0Xa_ )((\h j8t\VÀE85iatH\)&ΛG?Kq`|)4| ŵTI3K0ӿ棌ON:kŵ`*niƏ>CS l~Zq KHߙRV~,F]ԷNr}.//g+Q^ⲞYfvQi26\lj{Zד`k{b|uOW!kaXfU@ }*3f1~v:ѣ>Wc^ z nrgRlBJH}0tW,|bP/,iQ ֍uu/g޼}.<{}~}9زos~5:bb֑$|؉w ^x]Sko5Ul9f7˜C^ wS]ڮQiu߿\J};= \藻5Pq7]I]}᧏uz׷;*\DEߗLidw+ TH}#iƞ(U6P[8Ayiv0 FEE:-:2댉U@K{:}:F<݈!ൢJho!a:X!\N e1K054ZxecgҍdO7Y,Ȣ]SH0 s=iVRay3PJt=C\di\] %TJP_JHn"l{6y3ϑ[}Id%-S(((#D3ˍU*ZMv,HLv kR:*\$m=G}IJ&n<_[niZ̡~Ғ Bfi,QQ[ʨDhU*k3q4VW PUجøŮ/])FuέvgI8 &uD%ŲdFR˴2aLHfOn~jӭVM0۵WI{7<+mkiŰRɃI%PR$j+J۔+`RA+%8 uׇ˾J [u-+Ueکaz% gTFˮ.rIuE1%sXRAs\6/ڻ Gz//"aQ#-LyG4F%2pj36򠍕RuV Œeݳ!D!eeZ$X D佖豉hj4BZ"2f#gU>mߪc\?gЈS *(Ёm8Lp5H1Dib4I g' qb%*Ű܃A.V5\ EI ݾޖnp}/d{s{zW<ȭ9uH<9<#-%1MY{MvU+oP^ e%|KvJjRݎ\SHq7W)Da I*"К%R6Xc* D UcSE#iH`x6N0+D{!1.DʬƶqǶZgڼڭ`V"y$".=F $7+U^ba)EfZV}Y1 !fFk|@N@`8 #QiLf}aI1FFl5AhDąaJ=! 18R$K}v(rʸ*8vY5V!!`: > " R9H0IrgF2Ge#\%EY/V/z!cGdƀRe]r7qbabvb#e#H#)AR{ټ]- pfPsp:c5.VH=Xb;.J=RPV>Ǜn#_bA2EpS[%G;}pQ@nT?mcgỤG-gnrO\.:u^wV&쓞pqfԻJ~w̿~uD)XeW;{\*3}$1lw-|;g.{>Pnv3npw5?H;)nK@'4Fb,˨?0ϵ /yx\ϩ8p3lG2Fnv;R3q77j%YO4-lwǬ=LŸ1"_>nC2mhl߯?6 zs_^Emއk7=kV*k- ͞' jG~EH˞۱!ԙTJ-s)p%#njۃ>K`kgd݃vFkћhhG,-6RJd?7ݻTkԻs\QNi$Z5msS9"{\:Bei`' 6ײYSkٲj1u.9+j/@Fk$M|߱VW.Y׀TRTpe@YC&SdKU0 3K>B1nјUr=Ωl5{ ߨغGJ<8|AF]K*wv{u=dC̬ ; :dyAL%;K}#}PRېY:Vfέ4jh%uVRT{tFi}AFszIɷ5r'ؒ38:,$kc0eYh3!0,E&fVV 1֨|S!'C;X+ܜ #wS[K1t5XY?ta0wTL Ajim9IBb9m!% `N vDk ^J (~Nٍ-M=n5XBCGvbqgDj\PP(qFyŒ|k0Bb|mh W!CBl,@%r$F+=(aLl!{k?VBzu }p@o>bA^<16900ƌ6 )`:i "0˨s#w [϶3tK@cS)8c&4*` K <ѣy _vhe@^xoAZ-PMf,t^Øel #.k: {K̨%+&F"- < 14j_`ua(j@F5s%*S͚:m`ق5Hܩ`M0>O7"$7]eS0+pmm,#{ Х$~ PnF'.R:5Bu@ @ R'X1zhpSh⬔`NJ@Ak^BZvׂbrC1hwVX2'sľ .Fl0Vn7Kc1ˑ ;Pf hҙŒH jErlAHҨ5A]B}@E(G`j?&lzB^ǵ0$va얗XR 1Uwǝ"bK%U 6[)/ܴA^g߶!~M>JLtAOQ>>ߑ:X }q9U.wVi[,@RJjKPrѵ;n?<)#3?;#ƲJ?ٗԊ ,CĵyYVl2鏏9vK):1(r@$> #^ cГq&/ I{C9|ŕQ==尠3?tMn+_r|P5g&9slZ٥4ŇcSQTU1j)pf4=xod9Mf:[ɻ+o6skbO7<{Dd~C*Ij:j-7tv%Rv9i͋@hY==tXy5p yW &Q//QטKLN*6C$oKd*;FU9+FSl9>R/m>NGl xQ VWB2 7KޔH;q{,N{ǜVuBaPo'OYn0o]no>~V&vC77NoH3yc~-3ߤ߇zv}Q5˘?kROF NAGMʫ)")")")")")")")")")")")")")")")")")")")")")")"VbDpPF-Fz-"hĊ:uDD=Eߑ(pfrO}7uἄ__(=!Ҽh,P?XkSCQZc*ݷkPD RwXƋ•GpyƇl3&TM#eJ5rRwUW :6ϼ‡sX˧w/ӂ߾D?g_KEvS]m]x2o}w}u?tUs?1SdTOTOTOTOTOTOTOTOTOTOTOTOTOTOTOTOTOTOTOTOTOTOTO\'5v\'o(:u(m'\'@'5ߣr1ޟ7w?Èz_ְȃ%/%<S~"u:!:`5ը PM)Ѻ.70l%ǡ ևsXyvWOIN/_<[."_\-7>Ϗ8CK_o"ֳ#)m=pm>ԗWsbVtA Fn%* oj9/' ']fG/8zsrrwc8/ngehnN#lߝ'ߒ#p}y|wxcrhp?ۓ8q͠%qUo4ciqia~N`ױbMo(&xn-v={zbp@d&gX*ixh5κnɗ qw{uYr=ŸFDY][oȑ+<%Ah2w Ζ3~ ;XsKo} .Κ]^>T5+v1-[qÚ,?m/YVA43<`Q{lQ*}4$HT1Z@i`/!{[}>{ ^kg}u{|,U<<>•Wr GE?ǢXc,E?ǢXc,E?ǢXc,E?ǢXc,E?ǢXc,E?ǢX?MBQj_'jR\L<1yfs YNrX|*UC W))r)W _2L|z0 dZ0XvQ 1Vr*uE^$"{!I^I!6&D(Sh>D'8ӵo"W#ފH`h78CW/:yV#6p=vY?B:*5sk&- il8$hބv&}8OG Y:i(crn3O6ѓuz\2#`\]:fq2pc|=!`,$oKn~icj2}=oqƭ1;ېz;/~ex>+|xegmE܆ہs8 |şgy?^|ދl)>Mp1U6qvM_LL=3^ ڟV`F `05gu(a )`x1|hXpOiQ«F:>(0l4`H8Y6Iۯl=zq0 8q6M,s = &> vK=/a9]O>4HWhjXC9'2!Zd%86 jw)Lʣ UY,ޥIO:% z>tt^X:77XE!PIU+)/1S4 88ESWUTQ/O#98toD'HMށQM DSehQ6^RJƒXO[Yہõ-fP2l78x 7b dqhq}1W S@}_U闙0J(# M ,ꪆEV NBe5V|(YqοKq|O{xf3_ͳOpw)<\'w0JI5I%a$h&~)4ێ*Uu1PX%PMPnbm$JΉHZ^yA;uAF(ط6{s;omHX!D4o=?^FS,EDjoVKyP# MiX1_ sy9 1bXxG`q(_G3&2̾:n, 9*e-Udn5q,!ِ8sKzzP;WeDieK<λ9&W BQj4,; kM;D꼖"h<¦xPg#lu ~"gx2lZZKi"0[+C FiZs D7=*3̝f*'jЪ U0jl`~Yc}Pvwror16v繽 -41 sCzv'Zs]'l^G66͈c[mlm|6Iܘ4Noi ]5Na+'Ch=?0&|w^LCXY,X|pd=_^s'}  O UyA˫>O>\[;^1_L3oou/ztח_ޢ7#>%:ktfUe!rRO vD72pDιKa=)kRkjѫ'>6!re%X>O,+H~z8?9|uǪZE),XA=¡4~5T3 dCy ]NBz@\TxvOw g67^ɖefojnn"(vl&Uy(axRZ)3ꭊϦ**ުVzoUD8[,UOT ۻx͗kn+@V߿~ņ6E+hUՔ- %7FpXj^-ѫtH /vKʓ$p#~]VݑyvEȷ{<8Džnb 2pKBccSF,7i"xݸ^g}mBx=d>AoY/%2Ph%,s߃J~6V |A(k>4Ut]ֹyc}=U)> =5< f׳/Tb9ݭ7U`07Rz/d Y9B0΍f9J1 1ۛ'rj2Or L}LZ(eSUaP}'{o[ܭ.NWNgDZ?X0;1w7{ swCMZܲ1,;=yy_eѼ7+z Jؓg8F5|M'&P<Hc:^gj-~T-BeB"\L?6Y ?o@1oY2£GBk<_۪enpU7:1RJu詵 #+kD.th%d 'L[]`D6tp+@k9MJЕ `ی # EWW\ *:]!J]S+<$#XnO l | QrYJQ9͈0r-ʆ: ]i*yi ]!\r+@+IvDI3xte6' 3pu6 ֒ADX+ws:l l+D{dЕraG>N%:\60]uC+n(Ub֕@WCo5&ians#w`t8&*-e#E>qYN#ˬ)#a_N,V^ gԊ(K('݊e{L0FCvCkl.6(.6B]!`C+ȅP:PJSsÍȈ`&BBWv(n(/tut% \t  ]!\)s+Det(mqOJٌ k+;tnhI ҕVS]`C.&vhҹJKɈdZ ]ZM]!J^HhN'XS ]!\M5NW25YBWG+96`%Id@&=x Q&zXj2[9# :6rJnp:d J3h:Е)tSe+ʷ/]`T6tpȅ&4  ä̆.#*MJ ] ]qDҝKF+;tV{7JNWr%QtJh+|־ j ]!ZR+@ -tut% U*'BTBW'BĮNȌJA2:D:gҡLaGY%2`]!|NJ޺B:A2J3+lq12{+D˒]!ʕ9[tZɬɈ0W$2 \M"JJY ]}[=#V)bqV qVan>^[.JweRWF9x ^.TSwd{2)6؞j(D _+ 0CG`tB@=1LMvpmqwzF9#0 ]!\s+DLt(S(tu(@3+͆ :]!J ] ] YNtM>µ: QrSJR.=8µ8VB8HWjItFtWט\ Ъ,+(99X|bWf2}gQ:E2Lru'vpY6 4uBUb: ]Yf$glLDZ,tEvCOϐ:5Þ vCk0Qdy(ٞe)em3Dz&(m>ގ~Fromdoclz}gM=2; oCC@Ɍ "ܣŴԙQJQqsV6C7BBNWRBW'HW\Pks`Dl * ]ZJdt(+tut%ReDWB("t6tpu6a7@K! ]bxn$U8 QZYJI.wm[ y1Pi t?֍,l'- Eɴ,YDEŦpbq4|fgs8]`Hg+tѪ]JT)ҕ.W6iWʣ=In Q>wutVK r-ЕGˎvZIgi:U잙EgqfF2F?iO5 ,S5N #ZeY2-9㕾™R;gӻ*mܙt—j\󾚦ueؔS.8yӧ-W.iRqY.$+g-$f)dU6e]uJUw\Sm)ZLOuhiz] WO"ٓ+Z\WmߧótIQY)T}5§ 6zDók<`/rk;>,hA{ȯκ WBKڳ =G~G_N .7|ÞM "zѸc{7xJ[Z [|ar3"f 0 asx@(oCK\ٸY5o^8cʼ~;|yB8n*Ŀ7`T}߯~GZz۱ö0Zv1^^y㽾hUld8_}$;)ר7]̴- b08jCϞ Kd #A rx6r)cddo{ЇɾmmkT'kxeTU~_]X+M@}5K`KfVetfg+Zĵ[β?6R@.$ -(VFy.?X=ΟSԙ`We'4BC FXu'p7hcO0z L0-]yJt.NtrtE%'RW莺pUg;]ytutŤ"Lw<`ЕGKӕGIeOW'HW\A0%u~8Pg5C)tOW'HWB=8 ;CWLʣ PRJ{:A!%;jЕGKӕG٧Otɠ|mպ+thu6+Cϝ9Q?³@h͍ӋȊW-tF+|'4?_5w0&>'1I6FTG3Ǽէ%LC!6ֲ;!IWBlL%Cl>>0Ct+) ]NW랮N(chЕ{weH;]yqtŘ$G&f#ЕKqWʣʣ+T s$:CW.]+c+NLt)v<\uJutQꞮN$vkXЕ;P\4CIT)ҕi!gc7{IӕGuOW'HWZ'W[+ ]y3G+~ݕGzu|\ ΑHn5Co4r7ыβ2?e7nqʛ4. ]֏˹/61lFÿ\Dr DEw *MW.Ei3o\^6L/SKix)rIBHDpG3JE"h k`Eޘtp&Ф]Aj'-PiN?..#Be'b$420⌢ 1 TZ{`8({M6Vq.`GQ`R.aWoc?(t#|!MG7'滋]ޘբIzѠM2/S5ݚl0Fr]] PX)C Wi>}fQkr:DnVڪuG_n週剡 (Is0#Tk?N2-F i(a2.F؊q]r@yg7U'`wU|eRkZ.}O=ͮf( 5%jM_?[~↞X=Cx ^f4ڛ,sEPjqkF#}ܹ*M?E o%»pW_̫SBEW$D#/dtqқw徢(va9.>q/04BkhYg7 nAB B_S],΂͟mr$Quh]VAS*xV<\e7 Gq0ʍ)ly<6trsȺͯn&Y! l&O=~&΀TUƼ ZI9FL *y@mRizr*%Ձ$D-Iގ: b'> pb yd㐉ąU,$qd"EK},\P …|,\_͵ S@Q0osNbbp4B,rqB$0 #ͬ F^VEKe},F]qaZ= `_N>xL\}eѯ]Q_}Wޖ~kB=<^;v|yWQE`BF -J01bVE䈖I-q"an,5ONYJ5>[W!(j;F߻%{. =.yj?/=_X$[&6k+8&F KJ\$1SX #cK-h|w[Kj/^8Nx siqϿo~z<~iXfoPmdup9^D6r˴2 &/6aR!w|̓[<qf.3LWW Q~.Nc>y?A=ۈJ(F+3G5/X+.O|b+)x]DT/_WlHVIVR#yra/ŹORAөW~ЏnJCrG5].VE5zB;lkϚ5(ּʵo^-D6^tU0b5cQs0ogX0ޢO.Wlb ,ڶgDY-)F6FURn?d<ﰚwyI`|OΗ.9>^ǺYuLɯ'M'?⟓6(^l[.>d?SwT o=ccrrW{PUzQn_X|v(67Z:7I6-`uL]3$;x=Ae؇nCeP&e~gf >'n4U4vS)`ރh#{ggΔ=79gVDyqdC2H*G)eQδGB+,k#IQ>]YoG+cGdFzj</10<"^7i6iJ^Ⱦxuh&؂$6+"jY>ٲ E5r5pݥ/Ђ,ǂ-*fIDHDI LAsa+ &BR`*VkFfv{;­U[[o6KTo'?S6Rφ>:`GoE=i xʽ"zP k-ˇ~|觫2| ѫYa]>_'`SU ]U_Gݯ?_דIpF_&M'6<7un2$:dX.J:@&YWʸ n^Qe?خyN?OQp ʮV^Gf'mآ&g~z5{{!C^]sxd~{" )'AAb*b$_ \R]JW;yM6&m22ID,O .j06 Ō^E`t~SunS#S9 2XM"$r-&PKN)QV{(tRD|_u[(= uF]LҕD*:5h>$"eBr(&K ћ;hf1_M |b̷-1 kv6Hm5rD-Z1gQ|lHyPmP6]i՘{kyl/[:[s}EFO>[aD=`X{BN>ԛTm=`ܮ+}!nOnqˆI,XTW FVN֚1P1i |St4ئlG`ˇ5^t<cud+H"3)6 bA{/A' ٗڥܵ*9C]sv6>7gWT")bQu-;3ٮaxX˥ꯍ"+e{Q`d$(0Z4l6' >8:u=ggg|Q*"L1A`/ e$zlJQ#%aLE1 Jhy|7wYZ1}Փ!z zB,VԚ%DYIHmkrI-FAv^LhN -)DKt !\y0;.NjS9gCqpB4YbѺYե6b2+}c.rV#X:abMf ϲ6Fg h~Wu 3Adž125(1a0@)d*<#eM2l )2֗VG@ZVhg`# c3"n}v3型{.:(ޱ=W!vi0:u e7]2E+́ZbzUuɕ9J2 jVkʜOg`*\HBP $@Gsd_Iw[pR]ϫ`f_vB0M5ztDuncvN? VLiXq0*j6jҏgbo?ri4zBڑh0j0 a.ŎF㩌X5x4XŘ-׎.rݨiiKԁ 'lB59}ٍgkRlzS{)-M-ϳ F]`rzۿ}?wO?|ZgW+uS)QP1Ckjho1vp[>|1fc.zjRNHwݏg&Eѩ:.t&[Ń ./:„ERfZ|VvAlֻܹmXa %Wn(a61Q\ Tr b- m`J(<"@L!/u;Cw2ϟ۴Hc:SєFJFOQP))uQB},u%J2;NV7oF7+[ oyc;_FFC/q] abie&Yb(^˞dtcԻ1 ͺnm[6ےu\^EfS& u2ԦȖSdlbo 9Jzd̞1Y )e[Gr BE6"ZK5k5\d&=Atza? 1q?8UkRR|Vd|d e){rh2V d:Y`$&G,18k1*5BCpY>z+w$/mb^ e R4 JШ^H&d5VYlxN腶)bW6|l:۵s)P‘QtYdmaE[ E6(GX% MAsA}9I=.2K^H-*z_/E3Rᘑ"V3N;;ig'=c88+R%(i1}X*4BrJgJU1ShmV9˚,[.#i][&r@#H)<90ꖉ)R/(fmvDxn5ִ̢=kF"2m|\{GZhJkXa",R-r,bQ,[s:9ȒeG;hK0֧mf)w`c׸ z65amɳF|I?_%LD ;Y$M>S :̓M/%ሗx.zϢj7kcއ\ZJ-7H~ڤ ̝q-is$c9)d#!( uNٛ(1 qVAס+@Hj\C_nmx\)t6Sʛ?Ш;2M.َgzߕ4yzZ1';ߩƓiJpK$T*ߧt$[fHF/O}UB,]}UVYUkz5jK™J墥bC%g0f(RjG PXHͦsdlUf,ԍe'g;=xM$gNɧI:N/> `rKlUvC@"f+FrPk]^j J' (Ne^JVUljBQm;<[̾DE$t%v-x/"QǦR4f'wD!&Ed*+&6"eCԐ$bMA 60Lk*RtF[Ȍ %$U\ X!bjsup_-F["N"$⋐بKZg0TRA33{RMXYj(FQo*BXɀX:X9&dJ6 ]JlI+@Pw=nH@c`m4v vZ챽:R{w2dċARL7̜݈Ca\/is2//.qp[ן!kUjbMeSʭ/z5b'/,t*_GOM> _ĩ=<;(^v_Ek Ѽ,X=Qw~|G[(Ln ʛf_;OPYMyʅjC& N+ 'ZjrjƽDy#W)`Y Z7]K0j}hD+TJ5ǖvQOJfB6Utq-{ BldL_TX$fb89#0qv}kOI8`=:ߏ :[+poVߍo.>}j57>|mV],v.R'Bo!_ /s)e\=EGwwBm q+KyGzǝo\f[vn>>su7.-xwZ圅+\7}9\p'L9gY=PnmLGq= i|Bl&A@6!d+S}>h%@N/HӋ%$nVS-IpBɈֹ)%зh, SfC2 C& d9{jYN2\?Iʱ\̎?ًvM=tKkndԴ'j&lZ!Bʛ`K}4jeԠJZ"+yyaHt5jbL:UK)ȒBo⛴ư8TM[-Mfȳ~8򽍝/x_M9nK)9+򙔥Nַ͗8t)ԣ c.&:4/.Zu>ǃz<(ESX7"3TsͶʌ{8V*Z9qԢiL$Za$kU- Jk;А{~^~zyvO@?%ǁb> !F~55huAh85!ûMq#J,jjCfogBsuHi##1T,%w.(Wl!JQWc!H AB6[ENӞ4o='8H{-wz3W,tg}^Ϻٿn}R^u]S7vv o?jWK~}>{dž8GQOg_d ~hY}&>a_f-Zf#olZ/ ˜?YR8[[\_gw^]2W>X"|i+WF mե܆_vU&:ϐ,:6]oX~_hyo\@Z{Dd| o񀣿Y-:31GNGƣ/gW}#˳ W@~g8qo,K5+> H6Z $GZN 1qvtFou]C"Ǝ'xtPzh, MB띍 1ӛvu5zZE槛=6[!=&zK0'n]Q+_Ѳ5G\q=R 06d{Q|A}(* ɵ;vGtϥ,)]sfb1Kr#\89d( ZáU-u1YYk⸹,|E_V+Xޚ^j;֮*x >u fWE.YRr%-JE\TCvW9&`YWKQ {=?)ժBZBn3Aɹ):Z45Kr0ٚ 8V;v?h-tL 7S,h-6-}#dz]/,%|cTL 5 &S)UZIR]&f c3ix38fcq}cJ#MBjMrJ]&h gH{}9SHuۘR}";7o!+K›Lw(ιa0 Ow]Y*Q6zpl>whcQ~Bm=BP:bvzrU{i}'I&Cm-Zi:d̉]Xş>B>U9jQ%ګ9S<ΩT"Ij6\޾EQZlt S%כ}[XՐR<PU,yKD C+pڸ\neU/f**FɂuŘQBz%!&D_fϰ|5ӽ\k{+G{M H|6-D"  d:@X8pi'`Ӧtd鳒)HكIrdhR}CNƲI;G* <@V [ .Kc4-  }$P֠Jx$~}Cfp6 mBI,TG7BK05Ww*V"BeUN;>n ƻhpONU=7q'>dѡ%MF)V X;KS<m=Ť:"Jy Zc @ aZt ((BoL(b{Ig[>%(ZVR;@r0]2| a%KȓrFNH˭?ڷL ̓pH"}R'H1~J6oO`9'[]( /!aBs/._^$Ei#yA0O$Fl`o40ZYM O#p:|tVf JqT3KEB3 @53akJMsߚ(ҿ]J)Ʀ"%YnE9ΏǬݤϮp2i>d=wX,IlKY]䡦prb݃cKq;e.}btE#W.B4W@i * Brf z09vKk`}QR'dZ$B0k%:ZnQѯeL.)Xc*KKe$ml*rUWY{ |/s3+2WY|]gw<4T#l!?A\=wgp^Oh0=/??:蟞zVUϮM N/pٵ2|t'6It6bmo.*@ sZ=#h8䎇Sn`wuYԊ/:.izY2at.M`jA#~{7X{Q5 .R ozu~~/_xW'˞'~|ߠ&%p \܄_wj;4ƭo-]W-n-vv{UWm?-GV|y0|bPB1sFx0cuW٠h[VieV9ք8 lnW\1b?[HW}[vQ@I}@3J68%LE$]B%TYUBzaaV}ܧY;Wh)z =<@0mȺϼ40,U>[c@?:k*']jtQRM&fokzrdm tg^̖I"[fm|-w>gnc+s L.' zd[n҃pgUpU!:W'UK٠&KS&!X&vEvQoC/l]Ɛ܇{-Qs oxp@rp `uw;g6{G`8ߕ_Ϸ6|2tDe2^@if,z|spYe Qʪy) @orJz-A&#HU:U.6;P;kJ/s PfԎ5O6IL ~8%D'~ioԥ1Nqvq Ů)v y BFw!w8e722('C/K6.=>o=-_cb3)Z"W!h<.;9n ᘆ̂{܂˷{tUs?цBdUƗ~⿭(d~l 70f9eIy`)WEakR.|]Xӷl՘Hs;ŌǗgG7(׸gp!XCj)<l>I\E三kzpYgR/}}Ѫ"d̲MpWUY 6n3CZ2Ap@*n sCexzQEJkɵ1z 4w*wLsJI']KDOpNIJ$8 )+Jr+WHBcM.$'d. s )FpUgg2$znv=miC!zc9VX,8ebxJ)$"E1gJUM݌ \=LD|c[dlI&(-5`6g t,F#lSh.OgMf Gwtƚ4/љ+k t[KoQY&paDȥ{߆Nn |o÷ ~[aП\l/G\t .\^jKZsfm#u@U[>嘲k wiͷhB-a7ָQve9!ߊkۨ]UM-8Y*urC{VF%KmCQFaw=.>Dk}q܃#1rubK ]c.JQq34!zTᢲ\Cޟ;kΖ&{hui}4jtb3-3FZ|Yr$ȡc^y8Q@A~;UԢrP}*& kD(,uc)ne\XQ;k (xK l#9y2ϻ\X^Ƕdx0^YyTwIƉy8J~0/D{{S;l{QF0&K_[rga!J~ w4j90lUr$P1lsB#յe9[|lagX[(;l'Modl-ܿ;8NxsNh`0z?'_Üe(ǬU+Yx/SOe4JiabK'n`/ddB 4dT4*gZH2x7Qr.[XtWwKZI$#ɖW=|2=(ykiTUEɷc֑JJ^Ɣ۱;q<ݙtjUV[VG[ց܀U"h2$93H[06˨ $ 7 y ujcyȄ 9P$,ja>ĬJ:cg=lybO"v&E-⫰L 3r)#1s/IIuΉ\X%ӀӡtȲ 5$:rƴ]ֆH~b.v&҉>99௸4 z#~|GԴ} 1J]ԲM Fک6(H}qjg}ې'(O"̑3ˈ#Z<ҰT8Md+Y,Y}cax2տY~^,u =['fV֚^6-.TlsFl_%_xitM/i:11m͖ i|oۦ /w/V|=(/.h _y+0dE=gw>݌K>]讷7XC-#-֫cխ+.S.{ RXc=^>ٮES3 ?EL)Gk5j$֚l #}|w8 ]^eNޛQ TU91%.Rk ōy*ʔW +m2ǓOozͮUw1{5;BɊ11JlVFkJȲ))m$&bH+Kno"QS FUˈqܧ""9z6EX2<7FZzR٭kqME^~6qO1u%7gB%H&y͠} 5MRV]x, YkbM`S.TgI{x"ޣ:!Q(k,IKD {ɵfe'2iw)ǜ~q[BJQ:"Js̐5BE 흱{W,)NuXזw"Miyny{ig$OG1_bjJ55Ӽx@94+ A<^.,\2W\*"d{Y?W_Ǻ[z %}`JY WFkt;dpȋ3YF qsݔ~*qozuYkw,4k}<+n,R5{6mX{BPJ8[y(G翇^ d⡦u֐֜ CSGx!6 = ;¨ b`"癧u.|*E1j6;2% EPP n9iL,XZ!pEcPZy3 QCou4щCq!wx_kvt?lwQjœ7ϽM׈0Q5܊̥d %^xѐc 䘐++o';Coo$~jv7T.Do :v::+(vA|7/8h"r|[5Kp=T]͛ZX+k,QL?F^ͫjs{5ټ*I6{)e") ,fCN 96%xRǨ80D:%t`!$sGV\̆USD_{21F es;yƌǡJɱAL⎛㝾NP-zѢy2 C/Q`ҙaL(hy,,ve/\t6C r\aHH <Y]'n8jN!t|=^%l/g9.YljH|Rph!UעVc :TJE >OvOq^{xS:ן2p2&GG˯jdv(G67w:&,E܃V02h (.Jl<׏˰etv>n;p_]Fh%\bSo-+~Z6ro>^;y\)tRy$mb8% NwߖGzNj=' Y7vЈ%7sT*.GO> C%͝39Hŕ5Y">3~Heu6A'wC\~Di$KAId!K~>"a^iiٽg|{x\|&)_'Mgt3ć| etwXc'.c< G9--v'qB¤PU3&MA]LS+.J 6\az%F[Յix`CtCV+gf1UDV|]aDW<")6t4IrfpBt(y1[%?R^l32D_!өLekqOnUX[eFT/kIiHlZ}>ZnAuuɆv9a)3[R-]Y n;XrϙFK ܿ=͖ZRLh,OhkaVG8ڴj֌5wd#t3m2/:vGӹl7T <2K.&h{^1q?Q,vɬnqMÞg$ZvݦXr ] ;|,Wl_xտ~*LF.dMRJH 8,ךK]SI/;lD7'q^MG:xu'?0#[Ѡ18č C3c7z"&ލm,[ŝV1sx5I~͸-׾o;ƅ'o^q#G$?<3_(Xg]n<`4ϟDƲ s)"IKs)kO+d,g}/=nl.Bgt*bf%fgvZ,rAI9*%SYy4 Hp(4Nw.z 45p>jO8޸c,ǔe\^fA+) n{gt$ CБwAI BȠL)n6*:=wR{}7?UE7 \7U>2p9NeQj}ա[Uߦh[?jƣmsřTfq_>kGUsҸ0kZwhq6zwwwҧ^5n4ziZDUճ [i)%ض0jL1}Dx:1uohe+=]P|HW DZeu<kRHb*JAhNOUs9!如jRP%}X}ɚj鉌jdzO^#1elsVhBqH@|BﮫV+2X܀\ >nl45i).зJ,Q¹|Zu;a"{+8#;YIhdٹd:MnYsKYv P0*YӿlƬZ }/$f{u /b s#j>|pKuS/_3z{i\jtNsN.M 3l1)x 0$s'"ai[}naQ{rWۋd~v+MMm(f_j106u2;=*is ޫBSҀ(5@` J89NA%)54vt@PpEI6 :୷A10]1댝=.'BXѿCMƐhXb0T)F%+=0 aZ:vWd%Nŗ$si eAg.V1]tYsVc3b!Ef5D>ȍNT`jn:vB:@bэA{.Ut4G6LlaIA&LbHt^NH&-}8Ŋٷkq.SVyam?ػ޶vWe /i8/~iqPv>'pbٱ.i[^@l+Zŵ83fAvՅ,Ok'E*>}3s1cyHBAu>s]edI t 2$%)@bAC"w)(*hW*KH EImUW9@Vz2!Wo)`RɧHpjXNGD M fGl!Es햾 v:Ƞ Y<#f_!K,ΐ-X .{@ي&Qu$kOAC$uI}:x.Ke;;$ٷzR2sy Yφ~6>_qm f^=)~usNAOer~ >MnOtrxʰa2i~ȃ>4}izq;ݰh,}9Ws\gn;vVo5Xjm.# ž*|F`KHus}v>KQ|8m_YF: \ZQ0wop^^]|9evT~\,P0$1X"UшbSi.HF7JJFm)xaO^6#P=ELAb$1݋kneM{@/dM.Ƈ[b>)c0dO$0iszkQ*uM?\|!E.D[3v.էFf5}͝n|͙v~~ƶ9qosfxMe\?"|Cݭ94Iuel1V^ZB%0BQp<ب1Ac_X5.5iG*sxf] ])U{S)@D<@QJ2T)Y_m8{Bm5qH>_/:֔,mh!aj_KՆ1D$oM t!io 0pZe !ר+*Qg $V ;&$!ށ[QrՊƁ۱Qu^jz}m@1>}2L^׋ʻv*Pvk EV#tʻˈ+dl]C.PcRhgε.UZF]%)v`jfSUV5|]!J!Fe6ݖVRK2-4-|V[Q뛆7F_Vws7;"I?o ;1E!F|uF&*6HZV!O*(cA7rDl_xM~ ًII̦.[H@o'}`9Rev6;]Li<M:jhG( C*ѱ J2oj2\I1ʹbd62Lj+Zs=dF ؋ P} Kc1QE;nm~YccE|QMfe Z؅UH ٪LȲLʗl'ajWRZBNb?6;(e؁9'O"vnx{ZxřX(1uvClhGMWx%DXede@g_.VO]t?Grk(]|x,t=<3U};ͮaܹ&P 7F?RwTGUj KE d$lH)K,V6?49?6rvC~X쎔T,0,sPZWUI[}U6 _Cn i:e<.lsdhd}1T| R.r !A(F* 6=؆Rח#pVU yNja|(T$}d fRQf ;k56GWZ![hGXğhqx5Qsi2igKEh zb'ޅt"BTb# cP6KZjTEYqhbA.C ~wRWA\+zjΧg݆ z '/MU#}&%o2kŋC&вd/C+zZKhiIܼ0U F8[dWZAxe`$CYnG5@^RfVdjdzɪ3RŰd)%*oGvsU}49JDP峝oôzdiA^ɑزdfȔ|Jg5ۦ\*&9O=V`Ȳ;Iȡ:k#nš`n"&bJ[@š,` 6y@B{x5(f_osy{T ݋tk*I*ȪmT|:3cElQ0vK⩟a$ Ɋ B*q<$iѽ?Bļu |xK[o\X2/wGHo):}:^_'LY+вM(;9]_}8MNxDzl"%0\ F*(WF<&7-A{^ZV<"ZV< ZV)c;'.B4Yb+CYN4Y$EP2u cXIHaYڀA!D@JYzhBy~G]>&x%ZgkEօ1)%n _Lm~;|o8>WX1#^}%eE—r&Y 0UkP BaS#8`*r)C͌H36kk,/|-fMsXRKJ>VX@գayelHxz u?*AnڬC` m>0'׆6o)skożqNg;:]irK-Yw]ݝ/Nyvyv_/ݟ]o YM y[n#yQU褿 ํ+FB,Ճk8Gp;#dM-1Ic1UJUJh|+7n*)kɐ`|+jOTb5mDo$m>*Cg߽pݯR3ٗ ~fŢ(n43xs}:o_Ϻ`ـjF]k)H!Z%gKAL.^R*J 9H)UsP8 $!DK#))sG[آ E֘T53݆[Y@^7Y=m>ܕW+y|/MgĖEvI|dgv!E$`ERDuJ8zc(jBUVCI%J5ed]Yb؇Wcʁ*l= Bז3h)/ ?| O뻐c1-qEm֢\j >ٓ(kV[U@A$LAT NvJ ~Ve5wdhr{v}: B~Ffeuoi9 Lv!Z-\93Etzi@ ؋_+Ր l36Mke=6ec|2+tp^LVO_ݜ`S4iN][oyM2P|L`MS?ߧ 5mX>&{vq;Ahxn?e˃28ܼ& Mhnٮ{Z!ŕH¹Hp^<9{@=A{wŗK^$xJ4"UR;~U҈yib/aҔdY Eȶ6XrF2*HY͠F+Ĭ1r\T9m1XejD5fZLcxuqd{hP|;"%<2vVA4~m_. 0hS^We0j%:A{"k1aF;*o]w?@;d 1brbݝ߼ȑ^}N@<;{ƑAp FG:ddwAE*$e9~ç(ESC$vTU=N (8\YQk.ըږN"*?m)8$F %wP (*M (B|CtךROmZ Θ r- dSguyکX= 7ѭ({=:Bמq mQN|+t x yԗ[pjQB'TR+mTO2.slg @'bZR8J)k(YpE,D#%<p'_8zq_̵g͛QD-&ӿ#AKN訣4XԂgKٖmYiцP:$s#8^mvyNZ̮yuѭ f,|WKKw˛/@#@5K@g|&&fP<%jT$VĒ |+"2-Oɸ `xRHၑ[ R;be8E,G_`Y@BX!mV#[}!Iv@ө"yE Y&w$B+uj=$W WǘO0JxO (s>-^B zԤe#5Co^È>)~(BuG]Y)GM-hݸL=h\Q7.gۍD2$:Z?$%j9A7'ccs1n_XOHz42BpěyL?.3- *a;zKrK;dg-/FP3U|b>Nα&!{rvtw1KX8=tmO;]F!aք쮗d ѼlZyC!jzJ|MڅM-ECyo5߰1#I{^Tu[ tI}uMA|35'/hzភ4m[:YA[57ݢY|0&lǼMd6=e!v5S,)yMXZǮ:'&6ݦZSuVD; N9\Ub'RR3NBKC"nķqӐ۩ <縷-wHxe0DI嘕.JHCA@HaCHTKzL>8zMrZSqcqTL4NẓsyU%(ooݓ0.}yb8{l!9z0Xf-ѳ˘K֬ :F;AZR0IC21)y\s#SM.:4A%3'bm^[UVy 4¬F;5&! D Z%q/it:pb'U~C..CNR}>3WvwS"%"usr;'%A @3xU ĀW{Qk;jJfo] mA j׳'SA1xU/F.8U#\Wrԙ\{ieЉɯ/7:w!xgqoEj,u>JC*hbf2`TIUuQ18C<*fTl?7tɨ$|"Q.q5 gM^}̀Kq`@l{E\w':_./|29ۀҀͷa&;睡oPY$SekG1]w{anJFpL5 ;INh$_3PP1qlGxz_Gcg.cϔf2㣌։,Oŝά]¯$??+E< 0'sq)Ǜ w.ĒUd݄32͛d]\9?|\leզuiS@+!$M$1SrIL ƭgfћ_}{=> ~CDmt-}k10W5cK~< *~ۆ{3U(e.*HPs/ :q[JZl2sP)c$p wxpV[8B" sA!s-gKݣ`BVrJR yDH$J+E yIs ;= Lr)/Ko3ZgBvQDf^J;OuRsE\ D3ɰl*댸k;UE b!H9IKHǘP534r!19EWׁ#xK3s'QAYܶ^si~oyrkD{WLQy2,X$nsvL)# 270͑\$.gJFɃҨx\`!iO*h"sp=⾖'Na^mZ? c.qJB%3c)d0\G#b :G^(mQefiL! 1:Xqd;P@@Ip-8$O,|:<(ll cZ &+RT.T"Mk2X?ݫs$ǣ~~|wO˞ӏ?Sw`= uK%0ܙl?5MMͦTw=-[w VW mO֊8?)~~&3ZΖřZq7,vECIMW%@;yRn0$)ݯ#_'^{Kxz ¹(:(,2ٹmE GL:LMGTzcf"C2: ۇ_v> ^Ny[$~#Ǔ/ۅC1KN}Vu?>~}P9~Ji:1`a\7`rLqӰgX0L=DLXEYEa xۻ>=CZ~kd[Z_Zz HLJt1DIM݁_#rCJ #C:U붴6+W7/>ܦG1eb\ݜha~4݊yQJ{gltq3Zdž"]u_s%Ҿ,̞ۚZ'a-ZD,Eǧ>@YA "BҢRm2s݆hbj@0g0Ec}"5kbPPRbG *2G͐ yDlA̿њ܅)]%bU*՜Iuޣ֝77Cڮ } Kk|;bOɫ.Ei j ֐.M1Y4XS#V$N5'd|i*Pc!VWHIUZ.2E!2Ύڭ;]1,+2N/Ju7;+ך}hruz$jalVxp%mRKB/L{MJQר 0xb$+.dKBE**XF3j HQT}ӌ\,M< Fi=TB ceWDj r&)e!UR#clJ?^o,4cbeڽM>7d&l4liq#=?Oy֞ ^2DвwrȌbԌsb*K-m^B-M6Av^G s$9bw/FxbJfFm9j!jv.'<C*g+Goهj1\)({22%u:JV3dF EFkdqt(#5#2V1Os<֝pJ-0JDfCD"⛈ҺljQ Xdٽ]VU#jbY*q]#b JJ灹1Y0*VA扵ř-92LZ7+(:*ݺ="ޜ?qq,uHkݬd߸;E7!.nx3Hj[ed$.ŧP[|ixiۑܺ!.nq(!+qW_kuejd mN>DLߑ'stwRc=wM<^z+I[#$`$s۞NR"~v~+"ܷVة>3JLd_soô{iYʴ>Β9jM\wTE,'c 4s' ;|g?"qXK:_MĴX伃WS P C$cx>QEʏJ5DžI4MI+:C0"3"3Ag^3;>NI(gŒjJ*K$j4VPP BqxF2QBSZax]LLZغRآeQYK:gw^ڻKls!IGc՗k1ɛ;pU-#xbՓ}^ 4;E|ǧ׏gS7#q XK Y[cUT<{ x+*e.XˢOLх\eh{5CCO)J#R_XBJ!+R-@uH)fs5(ЙZl|Hp͙ɧs?6|y /Yֺ3] `xPN03 fWV?*~X3FzZhъFt6:ǘ:`9 j }JX #0ԌMK /)c 6ݭG]uRN}U T=QyYI5O[.=wt?-NnڸζL mtY࿥6~I[ n=ZZo<h6Y7Rp˅4|xow򝖛bn撷$|=raLbt3W< !K\y(/m6/*jHHL_bn yͦy<G߲ݽ:DqEȦ'k Y;w6lTJTtڒ LT-TRZL&4|!xVUWAg΁ [Q!#jGU 4`HȐR3w֝UNE\^]|;#P?^^^;7+U=U#tdId[?5bu9 o,#ΛBAb&#ɣR)#QTŘTAі8*%ֶ8ةjtv^ށ7^םX~fk`jN1y,_YBdK,q4&cPHqiwpTp <PgzǕoU Th\4.% L2$iRo;\?$FrA_Ҡ+Ը%$Mte6ˆsB?'wSVrCMJ$EV[&F4p3ݟO>&Bo&`7^srznKc`ּ]) W F>0 ;Z6ך۷:yh.[T*͊0#56~ل*_UѲ >ú,.i@+֤goa@a)yYeX eaq1I +zĜn;}fɮOFR=! rQ$^ҳ\$:.7#% PI8fq%5<Dӎ9݀;J^W%P~m?UYlqQ:7ba F3(7a0.4z޶ _[irr }R3B',ItI85}׸]M"7ھ.E7귻zUQm}~u4ۿea :2lxф-~;1MALR{ʹa6ꝽA/`pLx**6;g\m؟{6Ľy(R6FyX6\F2|o<43ܣR5< &bf RL;̈`Ra!0X,2V}Ky+Ŏ:l 2(oՓLx39iH &41#aG, (Dxqp"V2Wz(ā.楿@ZԵ5޸uEF]gHUZ w^W=_+Rmq]Ie,xr휴Q*55_$P^v$S}))ex~ئm[l%շ؅uyZk|} hʍ%r5}mu״.!{.]|쵲2aDDP5; nUTsc6xΨjgokg;AikE{\&`0+~5"gLI(%VyLޥy<_b˵>k>"-x*Zϧ|fVvl2Hbփ95ԥ|K"iv2zgv V0a䐤X€/e;Q!(x^X+n'h{k䴑Y+fe`)g`.+넢"Ex"#wS23̓尴# ]HTE'c4q1a]KόW)t10:K:/VNo\L `d $֎($|>}-Z'd(dv/QGPTwLQ_u)3NlQ]XꭁTtS굄Kz S~b\:(epd$?4Iɺf60]=| vFY([Z&:s*|Q %fNRx8cܰT$,_1:~hRm-ATF60~D` ,l݀K?>oʬz dCoT0 =ye2 G+Q"Ű\ڭu+¥&Lt"ZFe)0P^jFl̵FΆ![@VJJdDЀ,q{ړ ƭ=c K鄁YLBVI|C|BB(*De.uXE0L-dG(o"X[Tn-W%c|{!cT[a0^"x$ɒ[ g) n b x`^bytG?'Ydm[N+UOt^ŵ}ȝqgQE.j~؜F-VqτVA3sCɑX[UjX}xX]9ڒqX`!FixԔs$kE[GR%"X:*eQ:$ͩ0˳]hs2Ikd0jM32uT-P¤!b#؊TNVk;ˋ4|ԥCs J._zEegV"B:I|~jdfYtb%O3AoI) 3ZJRs}Ľ$~>DIsk':(/2@Φo- xP["#ΘHiKOKKAYᎃwUB{C Q$ "vZ(Yz:x5xOw~ M<ͪ=ti;JnНNQ+6q C.R}4o2XOQ&i1=s:UUuYq.h[X>8(py$"kY"p7 ljĉսoӭwTDzRKEa˳0̦p]D/qG9j/_ %yzzC8jށ#n h/-.$"GjrX:iQQ:EG#]\E8.> i9?ˋ_46oXl|1M9(CyTD8=.>NJ^exq-~= zbia/̞~CqKv'vZNr]\R]Էe5},>l-j9{fr&pd'U߷~޶!I90hsk wLc1j&*^߲Leox0tkNYR3]X0z Ͷ7T--HTqG)9 0U*% 딏)Do--̫= /2#X.7xU+!J4ǬLu+f#8XhܸD$i@:@L1ЧrJv[]M5{|i=˅{ϫ^g=ψ~h>u{gf l<? 'SP} ̩RK'' 4Psω]W[wQ IXKz{/WggY-EnM,׳!Ww9zq}puƺuh8HKU60)yH9&agd+Xڵ|gVvjRR.ܓ? $ JҟaSR  ֝ZqUrJOWIolxe%)Ĩ5yy)Xh D@2o' d*Px_U-L{ٮP4lϊ#79[ev||1)pa\T ?A<Ҍu}38 CrpP~ߢ/f+Á^%lߧ}t=ǏmB'k{Vۮt^vd:>S(ʯZj>JC͍uz$2U:=KN/k JUEh_ET:p՗xuXϲը&c֎|aE '9?/SF3?&?ܼ Z/X)|}T<5lbFgtaJv[02pH52,vxI=<''"!bVUhٙ7|ttf# o[׈f?]GS-|Ч=uE{+ ލ#kw/ȨŎQ60e[I+܃"xsFwy\NN>. g|f Fi889,7]2۟/'u񞏧v4aBVк+-ZjZYFOH!5̻Ohx=jl0ါU}C XdVٸg>5y*\l >Ok\T.e9ߜT"N:l0^.ǏӇϿ?~OQ??~QovkE&-OFu:Uhsg8CCÛWŪ>tyBw$BtGs/ly>X*0AID_oA8Ӿ>8[3͑}&eN !Ř̹̔JgQJy.}b rPݧG`Ͽ hCk'ǵj=/mCڮ*/mPqLG$VAcQs#*28MK&!vJo\j/VFo{b~C;%7\Qo[ztW]5+m*LvX1d0줋d\9Y*T Ҁee!z{m*BjHa5{&U8B,9h ޺$N)55;#a(AѢ-D;\)rg(@ `T<j[#TɎez+'WoFVd'kx9cHCLI;tI&iƊi65 ܊(ګ!ۂSI%R@2>"L?BS!k T:.(8wjd+弥}G>|7s] k]{}zq&wޕN\clW3nQo^)N}ңP˰2hQNz_93^+r(0c$ \'J!$ YOkBBR 6X[J؄Lk8wsdJuB1/=/)/Cmr%bm;j|SfOu׳᠚rJD+" \8q!JcJr\#"u;3{b`GZ_ɑdY䶧QS4dU=EV=Hղ;ζ-:*_CubB!J%T֑hur2$Wط>P.o aӴbEk7jmXkQkv/Q M0wy!Z!,4sYmdTc]$4{Ȍ AM Xפ,,=Y$X}v16և͆>MiUCшǨG"40U!P*i w=-% E 2@ d'T#BX7F+:PR &$؁DIKTtI"h65○KgBtl]chG8};ްr42Y%JNr$PlCqu$n8Ņ^>l;,}V{bE6mk.mA\0> 1о$ jlq2t jE1ΓCV}F~hENǣ v l523Z > A#kX"~fSIBpKj[9E8tdsZCd>+fA YF)`d 0ptb:$߄eYmqt@b/7;Q|HkʆQP ;kai. i^YB&F+H6eQRً>+Z<c %h*gQT3:!A\bTPpmq٨x( uE BO"' ]D4@-$*MmYz"yH?AUmrˎ[.ivj|w@k-=-%=?x.:Za6ݛp\NаlsVvoEIs}}97WZwI#]0>_讎ƻw>_hݕtdIwZa<*(kW軏zA]$#JgGhIX0)*_-i4|ci^]]˺^~^ iEDF&Ow6>Af`!#KA11^%j gs6)^$a]HR*¦LL?`s!Dz`(ȣͬQ<nҐUhnl8KV 为eWًs9n)nb,ɳܳL<4f]]Ki'v!;kLl.F;{1@q2(ZVUM^e{NKT``YXQ+"}֘3R& 1'QAW\XtZ!xHDN"[k5'knCCPUNd=熟w'o!c1 Nr̶W9/ӍhG7u!mB>}PSu tD`4DB_UgKH:VGgRR4Z/lRI + I})BQ!xJ$^Aq5qn Mo}|[Ov >6%ys#{&dPocQm|_bʀ UJ΂5BxG=?x4 gP& @NQ-.f"(Y6ьRAHMhE$u+!9q(iMčFDJ!Rôfy0uЎΡ#w[?? [){Aқi?4xc͜/ h2,OZj·Rc 1HU6v&»@o$~j1AӾbT Aߎ37u̿Y_3|髩h7 #`;taDى?ђ@FEx5^;Wd/1sZiL"<')Ih=ǂXa9cSBk!!#wNiĻ̛Zb}_ķ=o6E< twyUw~4R֫[-/IL4A#7[Uyڼ¹uDS&.N3vݿA>m?`?|2?kZ<ֿ _z}6cJU}q>\-bhK[֮uYanozsm:_x?\O˧}'_~db@rRc&~) o'ŪP*{CP{_M`+2N+{MRGEVލE>DSUrdVո.R]T`D~rZrc|HݓVQhO7뿬2.^]UZU_^ǯ̍oG8kS]ݮћ:pȪ V^K2GqBc<ξxxG8]:O|xT#wۥeKt}2&;rJ~QQlQY&ތ$ZD $MB8!뢝I=t) ])Rʀ1*HhL>D( UCHVi*ަ`\2Hpz?Vt@QtnJD1;.7 r D9}X)DA8gBG^ )O82H: Irl4 g/hD^H/vؽ} HZ'/W#Z3pj)!ՙ _A<,_xhgH%9"<]+ȹ"跸j0'Vv)uږ!yɄ5UcɅ?iC+$*ߞËzOjYE11!@(IB^ EV'4>l:&YDmMAN3i$dpLrZ ǝ.WY#|K.RwVߛM_ /~j/LNZgP!q.CE9bAIF8F{yG0*gP  M*G(J ,+.(S, 1lلs 4պgUYPy?:C$(ke$n"t=֋Q4<GAq{fc)5R oҐTj@h YV"YJФFo&G!b?Z ԣ@Qk2mA6K-xE޵q$ep8@#?x! AG3E*ʲrgHE,ڢ턏ivTuWuwkGix»@2KEsI!yD >moBsGmBi5m1 a\QoLg{eڜ2U.օcul&PFu1%OwVS44(68#N(p@5f3(a 9%hRMx߿2bk `тAu6BrHiFtWI~Z?=H*l~Rf) ?l3R+F5jru vwG&?7Uz嬘8t݄l“>9kixtt }2Li$j]qZ2 A8p%D,4퍬I6jG]}ʽXf;7§rRHU#G6ANзgЏ6*^e"' R)CyGOWT2Ij۔׃޸zbJkbޅ⮡59cA@mN' p5}nWvuIʫ =E? oc^MLUwM辻]3dox4t{rǠDtqq8neƪ8Y̪ndAI :ʸksKayWRkj}8q^.oib 4b߫o.J\nh<8 &eJ,99yp+_!g > #RR-4RmIGo&qZmPl/xW]%|mwyS(P\$o W/UTWѸx[P¥~gqXT}8?PB_GvoGt|P ]^ p{.[\Y{h #X>:i^+| teߺKzv[%kՕ A(?e ռ; ,Ѽ(]֨\>gm.k9jU!r Q߹A,P4 'U1LDSQ}㺗bx-o!R¥Mu\:8@E|ĵ.qkܘ] >m1Ux4j0 }g[O)5äfVN{׽~l4;fNLoNE[V;I$_spbU\q5F{lM^*e6e?΀K 3:%<"Y[ɡFS\HN: =b[='uYH~׷Ow+Cy,+*3ݻ@b'Փ;"Ա8ԓ.c=Ybb$Ù3f̍S5 7/w=ZAGx"e٭W٣\IK4 E퓥{.<},b m~!#me%B @s4r%DtTgŀE9, &qShS`[m5Q>piԚ:qsTaT\*4R磶4EKX댜 oTTpIIZ(ij@b*\M@Y2 <)_`< </e\7X#ʲ,<1 ""1\ K2R ~Q\ՌQVdldLxT ,z(laP0 $ #>f8iފUIϾZبQ8~4X˩dcGJRPOK!iFV.Ԟhs3ЊOVܖdeQmBqh (D@ I!hHȨ:>x,s=ktr A~, |RD|6Jk+$U1B#a#y.ٶóh)mi5T82Xk(8P\f B<_c9 ~p0.֍JF};4{I\Q) HpC{/%Z5]Q/yWhA|KS:z v `YR紎.kvZ\1[> 8q19(F`9UjOrhM>}h 3p\$g$_F'ZbE-:A+)KckW8nr.K6\ ڱ˫Z#݉6xb<܍cbzc#:f!a=u4k>z|~rl/ݗ {r;;$94q Sۇ[AȪX9G\5 [9v0wY#a^p,|8N&ٻܜz娌urF]uK1 ש?L$$,}6u9(+cZ *VwCT!Ot4л@v~׿ٻ?.{/>~xn ǫHbg~މfw ͇榫q܈ ˻n'|qWD|lk;k-@?]y7U v6zdgll*@v8 b~&b!ܬJ  S:533X6Ҽ.(;I]7Xգc Tj$'*T*P"DiDi0t/l+|cᛄl p%G+PDz# 9/>`#КsjH38Gw1U0Cݱtxz]5~D#lRMʶ@ ;`gP9dPbϝp{aVcI#ø-³7&Rr})Uu19K10!!ߺ#MMq-_Bvw-TPXbPT- ڨ<QJOu|QWi[sz*4->\k:S5C7oA1vy3!B%qUzͳei0P2ZSNy9=2*?QIgڔ(fs[H-hlhZh)~Լ>i^J]0Wޠ*]hY[QK)pL>J'Az)ٻvR۽H!;3hY.g:Y>IkmPU2TEhdb#D/(I %ՉKYPSbglvާD;঺Uy|ϻxQHL&3 Jʂti.#yrD[~ȿ *5T$i^.c9\P3.ќTeMLXP;#g(3RXW'7@G|.Guօ.<էYܑ''xqr"?V^oF&#p* (2%km+GE}U|h,ӳFcΗ "{^Ik0}W[d)򵭤t"Y<WYjJ0"Uq0qv#p\e0,L3B ᄅ+Zf׍od?Kԝ~b} :pq>}mێ ^ E<rq1arRGu+R5޹ATahe %&8M̾mB9ʥ*a`L݈.y,=vv`Զ'>1{Sc#j[TehyVR|K>TNJD :O(0CbEP#\;SL`'ıR9P< 787ꗵ0 "'Dw`{fmR^n{?i9aiE.(:,4Zm,1D*YU+U16Q*6f(ID$_;Mg71BĐڥ{;ME3f=7?x}s.\ӑ͛<ѩ08AQt6'!.%G ڡJƂQr;cq_~ ~c9;셦Em nΥ1҉hT8" ASm>z8VpS`xUZvb Ŵg0-ȖDR/w:5'x /^z!mEFu{0}!b/m?R֝Zw^sߖ>:Wصxg*.w3N 4Y˦lqcqISo|&gI[n'l:>B23  `{0tǎCH s4d,ړ,cv;GaālMj!P2@CբaSUfrYd1MB`4mRA U>WBY9V8>8{D]M.?~Nܱe~y59nvM#N9wt=;㟒w>a*[Sk 2*o[YJr*Z[M>kN`DW.FT)hE%pE{c,; YT9g ȳu-7N}fbnJn+N3YѹG5SA>crvc`1!;gml.REN'd=Ad+v TJ[q9a%)1j4UVwB0%~U$VE[CU|ps'2=OM?g|OQuqC1QbeY.FBE.؄Xm,ւoW4Y.jlna/B CV;k@MV\e:p>;] pF۪Xߞ|nח ]|{KOadrGš1˙w.(!#RR(V$ZńFmk[:ŵL\+vw^笌V#%=ʧ2l@}8+Lڵ i_F?-c ~-yM6Λ_N?eKFL]S\-^8W΋aeDzvOg_袇ջ_4Jє9]~\M8ӸOz]Яr\E,gJW~xyHĐV܃ތEeDUwLN^̺9~Xg,>QgF쳴HnClruQ__UD ajӠt= ͏+²YHD~g}uk^ѠaVz~Sœ\т,S-?3s}%_0_F{ 5z;j1#i{7_>(Dmi#'_Œ;Sfu AGǠUyt!39Xoi0{77VTGu:\)׮% ͬ '8~ݶ&9fo5$뵘A.z`Ȗh\ -"X~o6ޠ6}#؋Fa^5IH.2n}]G)Ӄ]חbӽMLȫaqy%ϾS{pKAmNiZ,΀}8h=$:d+9-eۅmzϻ=`Qݪmsu}s)Rn*\I&tMgۇ5)ֵ߭U0:qTD!zp狷]Ŭ-m b];bd/?ɚ{4͈ދ*.gjӉ06\JlU##,BFf}>`XTôU^3nğgs1C7{1],}c  h $RDlucM(߾ysE+"@.Z/&%Qx9⏕cvw>R\$g*iV*rj"7z"Y0p\-I (`5]}G(E !Xc= g0-dիW2w:llmp0ZmرQZL@7C$( mVݻq+Ejs2|Jk1j-$aRxhTQ֤S.{.e1N6 Ǭ]Ħ֜ѡFFV%ʉH9k=@E1Q_D˺Qw'ao*@M&v(CFTea}CvD˪@'N,+~BM=BP:D D1͔QE<m]].6lAzZi/`{jAɘȺe?}0'3z*qs*^TљLLzk*0,VF(ke[c$mӲUgC ;DڍzNà"3TS,cFU +U9Vr@3hr :Zjd 1;*v*ZLs%']eߎ0QeLTgVhM. ¢O(!xj욍X"B{k: 9U sP3` F~P5D0l剮Z\nC)_d/9B3jڪ (oVbOYEƸdhQVC:V's=e\uLF ANjK j,J"cʃcF<@ E&rZVȼ"a|R(T)p|ʢuxq `|QHVǃVH Ef06(vd! T?`yrx*ƝfE 8-&sJ uYI N;o[jz~/{!O‚.&ڐ"b)VX;KnC9(P"Q1A݅ZZt -H Hz3EcY}/4Kq Ў2ƪvp :tѫR%lP}ƪ*ڜ׃jF4z,nV嶍JN;JD td."ΠAb&Ybu4$?y2ZF vQTDhQÕ,cQy7 s.1xXːBPuDYllsC L:C:hˢ9K4YTJ Ѐ,*yQ Uu}YA܋U]i^! ZH̤A@ Ht,2 c%`&kҘgoVgSf|Z UMv6NM52 . B][4q [-d0LB&Sns`*]ӿsFEuA$<m (GjFCk,յ<=7Tf vcQD[9DKh wzB # RT", wBA)҃0V6bXhlFT-lE^H"\x>\N.BNa|;?$ ^ѭ bS!'BԲ(0RXFUTk zt ~ aQЏ'#j hNiu)Fnd_'5֠Yru(A I *t$Ơ~B1X߲Fra\[ἨSfC}*fD88F7APX?˙P#,+QZi6C >bMtxo*[n7jgC1IĪتnj"|٥#jP "!2.Xr褀)`$din VkT}"!(Tv""DoaL`j?.zՅOz-nN:V*5cT(M 7ńhE^%x_/l=qGA15!H K nmv$-=]Vx(ulbS"`wdH cTHjNU+ +$w^+""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H "zI Tڞ $@W k?z>w}$q""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H "zI ONz'UCɺI Xi%@_# ;"H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H "I (OH kɐ@'`gD}$EEd"H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""!-GEK6x=zˤ.5w-/ Nb(M'=!t7~Uϫ}uh&9N?uu9>3 \OEu,Ga4)G/߼XەKqx;lYzWf6Jur>P}}է}d6=/{ Z},ֺ6NGڄTҼۧuspe|O7sq}asc ɡ\L'o^p3/}¥ԫ,A*.AhfOR8o%櫥@\hrKÅz&o^l5O16(Qۚ:H|R'Ki›"%7@*&h4PV9ʓSo c聯u[T]<  D%DF;̷tzFe ʬبMm*P ݞ{_UTAw`nl㍨G^N=Iw_ɑ+ ~E!i䛧(ge< gT,wܝ,<7ovI_–&lʼnc"n{7 @vzБ>MZTa>prݘAk9 5q+;O:g-sz53[wlnk/'/&/sΐ9um|~dz-Tfbַ0(ȷ|`Oh|}[{﷟{%=f=Zwwxnt=B^2ɗd>{ݷjmq#w67?u'q9>VW}7ؙ[cW76쟗g.خvmǓ jr}~7~W62X^.oAuw @LFsWGd\_8x= 5hSPGZ\Y~ݝFq׿+> JIuvNY5~4|#IH'KΈFǔe-({Et#~r#n!A^U3"E@iw@=Ipzc\۔ӝ7mȽt5nH.Q{,2Q⊮[i2r}CPut^%#fcӖ!Vq͹#ݣbHu7OwIiPPw%,o#,"տ]`y]=/V٥3(լm5g_>IS6JQUIʊikgRkDLҊm˽-{xOm?[;pomqS=3_q./<muM1QIw#tu<&U+ rKg3zzCE/-HdP&.XQ2aEA"O!W{?g'?]lUԁ YkoMxJ{^YƛXYo(&&øNWIҨlԯ'i Ңg?U =dAhn6?ȁAAO/4//l9YG&R^X*_ ޶QjMmd)Na0޷ZOKҶ-H+ ihVחFaGX_Mȋ IrqiZ_2c{Yd؛nO18:,yXaX^̮Nkλ5|]"\w3Eߘr,w1_]jtgm Gɥ tʫE>Kt G?vI?pmVo4nRhljwYǟ1=Wufu. 'x.ղ0_fm Oo6~K~t7k\M31M:Y9*ھ[29`$L?y,oo)M,ݦ`*w;V'轛lbHiێ{w']Zjʡ7[xE㛾Տg_{2]ٻN/mh ~S#o-[/˿;8cNWpy5lRqЪd_^;3tFA_bfh;ClUDUq;,d/"Afą\E[_.UQR+!BrcWbm*m.fZ0  jn_aKnK.K[uǓGdӞYG <+dQ3[@nM}ss-ꮨA^ԾmɄ'F)|֔6sJ6jvC@'Ps`]x]mkwp5Ǽ#mDZ zI;R'hm~z2YD?|;<EoJ;0JC|M{b| m,Y+L>='b(`bN-0ϰOug=70wĒg*F9˚&%,YXPK7Ǽ SǦxEϤeK9}l;V ᜵ 2j+ V+?/|ɶyUщAX:ێgYiE4ɇlt޵u$B (ApTwU{ᇬ'Y`2A&їjD*GՇ],J2udQ ۢx}뫮KsѪf">)L!=}0r]NѠ<ȼ*rN#PMŅ,ˣ*zϷqo| ~=N~dGp"'/R}ul#n¨ˣZAolt'ӣ囿~;MOFFZYnhv|4]q۽+c}oFh=(ϣ8] +A nr<,8-b{㑼\72tb}emxә>ݤG{qRfi(f~u0xeq?z$U Ynz*6m_bv~i6z/lX^I78NU46Qy3e28/~2j$s3f_b"NP},߱?\Q!7oem`V0Msno.Y6yXvMp-; *FN/Gjۏى}>,dkٙ)oTsuSpLlk򾏦_wt) {ECfĖ#xG;I^hp?_^X蜝w%~PXb Z-K5iNɇlu2P<<46{-}Ѱ~Y^4&.eqyysS\v]^snb;UbdVD](Qc3 LQQbiTj,!A^6Z#udƨbE'(6QȐ"EJVL_XT u`еU``cn0rtܚ %"x*JѾ*;.@dln' ]0nm *otK΅L[r,9,;5՘FN&CUR"CXcp1wmU]km1Hb$ɦ$u%s$j{Hz,%3 ͉Ƞ;";(6VYygVY{s [yIמOځ G: >.-h {i:a+1vIx.U[a`=aPrX!-J\+FWu.gͪOJpzNk SʦP+w%~ө_1>3@ńe텛>7dWn jZȚnF6#刭eZ ,Apr㉍rV~r6ۈ<?aܵ>]X;@׏AWZp ۞|k}cj ^}s cŐ;ޟ ф܊ PKLk@FhBsKn2OdS18 Ŗb^U}Q   Ev[z?29wO Ym[Te.B/t1,QtqxX~%qg\>GXg->L*_rR()toUIaj΅\ˍ? CZ :5`oC5NbT`*C٦Y9sU6B Q\i]$Ǘ丵vݚrix*gbvd?ApTӜ9GTVyMe %ށXȦP/٠gl̀d7HCJ\-5G5 @f/@<.]z%(1zns$sJt,)lHq ]9V/? o_@ɶS$Te KXfSzXզ2;AV69Uو 3(W{CYމ* BEIzee%zIT#KW?MekO~+vdH4oG`j|2~1oVy߽lއ*SgydTE˕%yuYo}y24+L>Ck34ז?+ Ulr1+) >9 '<9;oo/?û_?7Ė}^{ pa ME wPS|˩-[/:+qKq^Yv?L_Ƴy2p?h-{cV(,?"deJ*~ObAlf OAWc˶g$x@^*x8 K L#E Lb٥VM((ʺǪJB1VkYnNz695-<2K K>hTrJX N1dX\uXE hkv:lUd`<#7>:sϿ*:Y y;Wڐ^i w"] X;tiMe(b#mXƆoO܍h> YQ&@r1 $ڐYņ\M{wMQ8=m^\la<}eA0M2"N3{2ڵT}K`3jC U0ToǕյpl[ͅ5Ttq{i#aWQ82L˓q7Yu5!qESJ aYS r\::@1*02 6w 98Ąt6yںvՐ$31ڦ"$X:Km-:%mȹ#`QeoK# k]tOfrUJPbX\M0uQU-(2iڋԐYhZ`*M֙a˯%+C ER!,W}L YZ&(FTǷ7q< ȧvH͙7-Y{ Sq %(#Bb䳧LJ&'֌.n`Z`úܰUǚN#T+qp <C.L)=[DXwa۱Ul:r VE簋&H*O6Tz{7Y9#re#]TV"m'&RHbf|ʡؔ6^^i]ݓ*տ&hcn0bJƇZ'Pk LEG?l}zt#1>YOi. , dպHP!.cI]A--S^&3v> 9yQ[%Rl#NXBXv5Ah; Nyʇۜbq:֜lq+2\C '@o4w4RS]1ȶJ ic>A189z7cN62뭥"]J;(R_Mmmruzېgzâm.ACFNƋQQ@ RA9mMd-&̞4kĂBbJA` 5& l/89w@MpC뢫uP(t~\{6ss($.[ZGK`Bi C%p5[_-A$wYڊYTlR e# X*pd:] gFkjP9FFfݚT0M9UbF>`ȹm_!|_U.oGZEll2b[^*CU;*aj|U?{FeП>._vzl2E$cO#y aaQY]đK>!//F%UQHG>Q᧟Y9*^, `"0ZLoI2zoV="hP/ώ<ȓ-Gb+(E C>P!4Uҭ",!#Gq3r0 Ĝ1N8#RD HDf,"eL5&`"5>;VI"ZhR Э0l`P-R)*6w6GYl<35;2#i}Íl_qj+dzSrhMaNHUB1䙒MfS`IɧC|CޖSjlvC:uB̵kĚYc-6#~-?}ouԛO>^L}]ë߶*w8,r=g_O ۏRZR} Sq>L.:ƭr $ֹC'Nd=R,>>I xIN.ѮÏ(d wo__ɾZn ޽ư_^?m)kC}X|ǃs,ϧO>r9[#ehmyҲ_G,%{v͕P},^^|/IݝV*[t O;+jm!sI7e: NMbMMVULAT )2U!Lڷ{3AKwt>srwE*l-iϤJjS.&eTHDYX>*}4q#w PDK!vnyn4m6a<.^],Ol{U˚*Uk)fNykW:[Ns"".KҜF`[6(ciE=J0}ųp89 [aTMFgoY?7^anb9łJ#|&f ƌEjb;uuAPf 3(dgB.mҐx *YVQI%vxF..ըUa݉rkv=mTZ:ĒE5&BRD`dsU)9uBW52M?@b| pݱ'2*bSo?͎ ':p|!3AL#v\VǍ1{v$dB6oG@&ʮH]2)tUvJ# Xq=:ŒkeY8kYgd)d$_+(dSwqq?yᘴ;,4Ż>-|Vlw9Q8TN"yʒ#4\,Fs6B!*ЌQJ"b*0n*t@Ns&WȲݳ 9#5Q*R+hd֪F^LM݋i[C] v~%lo%Io^w_ea٤wҫ*5s[`\d l5 أbKO-{27Gg0It6PFI-oߴ#:iϼg2ljN87 b^wERL&"Q8ոf5XcS4i_cIR*FFxJ "%0Lv_ %c()!!{Vҝ̋=Nqs>yAЌt h#[t{CO֢%TdmԔdfqL(SEs R.jZIQ4q:PpzCvFrJpEuE3Ա*Es`W溤8X3ia,y.^w)/'E@`">?D`I?o.|oiIڰI)s\:%+tvsؼʛl^|~Wu~^.>IE?w*ϼw_Je͵ Giw?{$cro֡^hZ,7~|<=;k='g`U[95Ngp]_v5׶Vt*ŋ& ċtnϺ [ݺv}"AƊ?^opv]fUc[oxhbäVQ{cXo+`n(/_ɚdћIU\ԶdUo>4FpEvGr-1,vxp?m'Hg3R3!>y,}JT稵2iȬ)PȤdkQZQ򾿞ryK˂T+f}]]bVIl/?-Zt\>ĥhU&'RF켶l2&i$ Xuc~ *DZb2ʸRVjeJT6} s'`KH}/{mʓS1oCuUiKwХZ᜔!pU5"K))bji{e]%BƌiJY mƬmĪ֜HC.;>{;4ZB)Q_lDޕˋe;Y*6=l} xɐF`[4v^8b67EͪX:QR%rX3FωPQf̶:xQmKOAdD5}dc1*m6ZL%3&WCHqZ_4$0%֦\i 6e4(,IZG']1Dv95dU&t˚?W\*am3a='K`, /`X:pa? >n8@* Gsc07AyE nXX$ @JNPA"1H&bZ&D^Ű2'E@aZ ’zOy 0|IE@V/Vsx^ĭ |`unT$ 8Usm_i 8٤W0dg fI66X"d9[ͦHIDXO! >]S>i~eabidQ!.Qa0X oݷO?0u]u2fmBMc`A̾ ]{ #ecwP{B^ @l .p_օE(! ["D(7A`,+壮:@FgGD ցqWc'H&hSd*qzTͨ= m.! Gtg8)ge]b:: g` `?ALj A "RF.'o0hFaMxה_$<hMZHo8FL,NE)n)8g1h,uܵhVEyE=sϱdBGl18hͽ7Uf xSpaR" i!9NޅP.7fhNqD[9*}֘2bw !0H:PVi 'Ox#rk = jD}z%'b]XD WT`q qrpl w䏺؈QЭ bp|(O((aQD`T!+8#.L y+0F݃@RaAĤЏFМ,5RFfV@jXzP.s%^d+1bF"1Jc?]-ܲFrA5mjEDD̗o:i m;H*ֈ›e֍:RbF4ᖕL-5֋#Ѓ<bM Ǎ0#1>d G=x{zcnmF>%|UhU]ڸb"&rCPn8I~B%^p)zUR7(cUHw*5#gØaAA;zDlfӕ M["vcu׈ńhEV~O#Uc핵 !{]K +".3t|i:j} )hmg'[O{_)hp۟uC}~)߀kǕ.X[s  ,j )ŊxZ#Q V$@P hgX`h pk} W#pU"+"+"+"+"+"+"+"+"+"+"+"bl+\< Wrl\}>A)W\pEW\pEW\pEW\pEW\pEW\pEW\pEW\pEW\pEW\pEW\pEW\pECWFu W_'\sQWZ;p+p9W+"+"+"+"+"+"+"+"+"+"+"+"a+ >%Js\\dpUuWPӜp9Wqf "+"+"+"+"+"+"+"+"+"+"+"+F RN0XiZ2z]C^ensxCjْ1bK&۞:}ȯ=X3o^`K*W0!F& Q&$3V),Z wζQzkΉL;XUTdzف >u(1_W1z{qÄz70/'ի}6;om_?YJi8-^|=Qok _p2GQkZ^ҧO05—m:Kt6ډvˡ ᖝxv:9,'YxU0 VggGw\ĕBhzn,M)|$vAFwi1A6t3ҰC8 p?u~{/v͚O$Z干OO;׷W,rf ߎ2ZVCՆZygjAm]rVDݥg^6.=KYvgr9"X#PG4]{iZpwN4uNM]VdQ+Rl F Xu%}U:G2hnv-1wUTv޼e6F:c{mV%\A\^v[_WĄ ߻XÞ}Itұʡe).wdXr<;%)~םSETIYԚ'W[|4WO휶bsw}ӕ T3Ncҕ/{{J{>>GOwWjžMR.YKl][k"Ak8NRWدG_\p.wӯ<;b:'Eb,r),8R$EjhEt7p7'J7S8vf(]*OW\4¼ް@orݺ({]K}ffzm-fFqfjÜ^vȯ/7MݽCTEM:+2yl.o,s0Fo,'ݦuMqB8-حry]?p77ׅۮUS;kVXڪ?X>K0^T?RdzڽT5jyknl6˧bo~Ov,V_'^ZITUR4M(-?[,B!]_Lye͏AD}IxgX\ࢰnkk(%d`1*&R)JP<VB!nC֊OlIu|,(_Ei\QPW%,/",O"ſޛ\םb~Nz8x<>]fC/hQ};iӯ8E}tn,E!9"K@ֈr^k+ۖ;ybV&F#< cr >cm193I*J7.C*{4<(u˸w?'O~7msPA{.dU^ vƙYgMq[Y?ڬ_)@~d,-6m¡ی(; ā`Hzi '^&pS6XBl?.KH[jP~_Y%~ϿjSbOXdº \3c)Y.KߓY%cK>LXzS};NMy}~}/_`T#}aax;6zoWE^o,W-+͓^zXZ:?dU:*_}FG?]E9ܾHY}mϧ6|p᪍Ylw;l>vC[|t|EViv756cFnÀ~; K\L71N39&i29'L|z8M*^@rcܕzs,?Q6̻G|A#=gxdCY'xM /G]moptr/ve7I( gn5_)jLm ﺥ=ظͶҾ]㐎{z ?Y>z_u|b opd+d~­8@@4DcNp3:Umj0LUyx"~*UW`~ <xnOc[SU}ݓ#wnq\TsfP cdjBG0O6i_߆Ѥox'y߈!K՝ Z a>Q#=Q'y sGirQ;&ꤒF94I2?և4' ns$اg=ƾ-Og{sootۊ6 ^FCmj焋<61J8:2QwqU,2+&{xr<&ڔs0].ok"5$Ȼ3B2˶1:`S^XwAKRUUiͭIEJqηz˽oBqWհV~AUgX&ǹ, Ug4U=^/v꧳ŇiLj{_.~V*ͪ0㊗YW'uլE{ϟgY"q4;tk4"l=wv]Ys9+z݇yyz7bz31Pnݞ ֥*I.QVʈnK"LX ՃSyH9-u~2~7/hYf.…{6AMַ8Gl1Ҳ3©2޼ D5SK'(*|GL>aumӈQ2{ɵ/:Qw)H8Yl?,P"Z,e%z~|"QVr YlEdRqjG{I^jp>l(?^ZܪbQQY, KAR+8スe*ߎǓӕǝ她[쫎#w-q)Cen#sGVC߫{%&b=@a`{N Ue $RkSN)Ue& RHɊ @. t6溑sO[hBȮ (2-d*E2%TN0ovh+2/q|0U6:nͱoRTc9D[UH%. a%VW4ܝ4.*wc $cc1Hb$ɦ$u%s$j{Hz,%3 ͉ȠHN-)yC6Uhv/mV;b:z*+yQeZy. \/,I*Wb;C4)DaU7J(±VHˋ8R׊`eg@SZBCFǔ)T`o:3V{:TLX^iY%"RkhC0OZ&`bꈓvQdu㉝r6~N9YmD^П1bq_ t=ҏ~)uQC;8EۊuhPnHLqh{ h]wBn@.>r?6EIP,cQPPRdWV];s42x0eQ2VY:#bn!M['ߎj;c_1'艬#zY &?;u )ItK`ä0RaA?zOޑხEPuͭ?X*sD,`)qV\r#S]6BM(׶+#st*(c'YE Eט08ַa(CVX-E$ضj}=Dgk$ 4D@"vXsA`o]o:}V!y|Ǎ"QSz%&1zn`$sJt,)lHq ]9V^/_xDj-jIsمm:,Yl4T9mL$H}iF_ؑ:Sq7哨2) DP^W\VbJDXrx=lo| Mo՚1ܘ͇qs05_&kwʕOe4[]yZGv&q_7|D"ylk9{kٛȳ /sӆA7-V2CjoL1\][$8VͲŬ*0+$\D:/̥$;n&xܾBl9"sZ~[kM啟ޜ}\}qv1dbz8;\Y.Ve|~(c<,!63ymӘӀ[*D ,t`[Ŷ~^/b얇[g5YmmJE9y-!mb'inCϹZ7O+NkZl1^xof{y飰w~x~h{ן>e?Fv? qa L_D  ReԖS|Ղc^=y˼7Xκ k||& .RrZ[iIVyBz9&ú(C}jCF%`d+CUU)ihU&n:r3n>K/z|լBw&t=,Q`Нk}4>cBAE;_: X iMe(M͏AtKcx*DKo2dE}VFDTS03TZ|dkCfS#fs5Q#CuLQ8Y;;xuN6\6_ڮbE4:hrY/ɂ Pf%Tk|[݅uL*/҃ ގ[&D5 iX?X)x&p~0;5Ad\v8@4f\8#!OeVGaD5Vld`j2߁4䘪b`ҡ<#tu9!EIfb. MEHHA6buUZt)Kmnުd}KsP-Ǻ|dV>(gq`\ U.$ l Z+AkPEoAM ^$B&o:3314Y">":DQ]P9w *.,HfqQ籜Yv㓝,7-;Ynz@ϋТ-A1Z+Vkouյ;Fjμ@/?w]W|(ABK'=%O 03MN[&!7t_w'Ԕ(jÙ29,UDD.T+D⵷Qksfg+M׋3vՅ..|T]=yxfٻ6yU_0,g߹6("Tb抉*Z Xr\CTtփi}NZd؋(M1{H P S1k2Rgݍ5v:9. Xv7UkZMZ{`ooT5BiRCBJa-^̕٪}3>c̉Abj4]SXeoPb8r#T1q$;nܭ~}RX4b74FT'[㚫V^auCnQlȪ: .iNrF فաzԶNMϤ2{Z$,HڈIw6F7rֈ}dyI|X ט|ݸdWox';V3dWji俚1V*a ZSێ#Ρ^\űn1}cCGPam$Qg|#U+q8rFG3Q3Q+y|wjۅNc4 2@zM(twTϡ6Vz<;"o{ő.B+WMζPk)$J1 3>PlW|7tq:=9R[ZLl6:VQ!$;or-y`5DĢ[%"H OVSK-H!K|u.R)TXRZ  NzŚ[#VkGm-ddK)Fgc`|<T]Těj?Ќ; Nyʇۜbq:֜lq+2\C du1~A h4ZM'tvcN i]`E-1wcN?cjMwǜMmP뭥z"CJ(R_mm_mdyںvEްh {:/FEbHH-_Xf 7!}[+L=!hGĔR4 AkLD/;J9@:&8D]/jpy z6fiUn=i׹9TQe-F%0Ɖ4Gdk/ [S$ }<+Hƈ!P6@GK Š},nTHFYT5Z402ִxT Fh TG3YκsO9_!|zX[U.ţ^"66TT-Vy0zE{ԪGMU_RO%# "r{yt$rBFsiU&S'4gr%#5GF řTz+[hb&c[uE.9jEOUnA*묠Et\+@RĞ5(? }8x~1FVf*Y:%:Amcp{2,nrAO!P Sl~)eegyjٲ `zDvibCk:7t~cv9{>/;i۷}rYŵJ#xg{HW$xiCy EBcIJ>_HmTU_~Ie͂Z"!YGPHE0O184I(c:9I}c\.7Mhk1aA%TŸrx@KUclWx"$QIU~7ۇW?y7{`Wn ]`;Wx w^Y#HyuÆcPmN_Kl65m6-z"#ZꡠÏ=\Q5'(O4H9fU/Wo :B( Q깶(i \/<,){ xG5Z%d6JlgKbfK{oC엘.2 L IzR?)N!iNXt$N1.h|5ᦥܺ[zݮm9!mSnr./z:e3![*e'«R\݌cIP$!A9q*)R׊+-Lma1TR8K k<.R g.GĘ#r- R;Y,ŠLtii"Uwn]AgVM77gyQU5hg~:+ާĿ1zp1 ιqO *@$n4 1{L2b:6oCo$W^ {Xdh9@;z_v[ czҾK?. 㴪).(f:'O5Φљ͢_Q4)g=eozdIS+N.")`Y ~_WOnM&^#U,77رH/&ujb]wk{ץ׷1r`kdz{=ﰿu7%wm{h|ELtaCK-ðvbͻN Ws=O΍Ł 8Z9iΖ?kn Ym?eZ$mE57o Tv|0,S 7} k(:J"Ep9u\IZ c-\6.]jˣV!Zr"&} HJXȔ-zU36qɃHZKaF,! fCE2NH3Ns/9x+DuL]öƦ-m|$acG}:vnzdZO~4nVa|L_/'mpc"xq#V!Gs6$V$kJm6")cJKNocsCSDǘv3%88[@b#(p1Gʑ6t8+㢾M_g_,6qN1Q%7fDoZLyvq'jB{4FT!V*@ a[u>RhR25 .LM S 5E(;AYR |CY`q.ZJ9,0g`/7Pyس܏=QB 8G-w"#R.~o.> [zļ >Z4%3y=t qj{)?f{I3L6&W=x_{ɱyig*>}z?cHơ(x(0A.rJB@mg0ߦwəZΞ-E̩_άR'g8ٜM.b~l!HȄUAr"bEdˊe,FcdDy ݾ&9*3Lf"#u$P*b `"`Y-Xk9t%/>?W3KRsysωBRDx1%y:gdJGV9kDNmjfRsݥ +W1]n>Vs8aoa9Yy~,۶ l5O>yeV^OӻƒH=Ѩ J-kuɽ]hxoT)`#cA2\{Q4ь\0>KyPYaf9阒4ibCAKTk4JBg/M_76/Q&~\{lWݐ!}7v.oi|W#z ƆbfMQ!ZaА>tbW~D07`\gD|Dv{겉rF!Btpk|f7HtVEh5汵hZ 6{MYJ<MHRlS8 !֭RҞ56.i)sC&-CFߗ6SV"-ycJڌn^Ýw|X;<+*nrV]ߞ.><'C([kt^j{~P?*)/}.?Uj@&uamuYw!|@1f.l)͔(c-UTGfGRR뵮%'!RcM0Rh*r Q˰&xE4"`-#5HdЙ!hGin#~iT:/4ªH:&WjǪ ;uXL?m MybqU? םno^1[:2 eI<5ߖFFOg0~yv0{R8zz?ܙUbo&RJT̀ $OEu(ɀ\-ďFpOFcW'udl-i=$VxH)LcMqC: . A61wǾOk[X |-J;|۴Z-*M?"Y崴(<4g]:FM 5x|&=ZYѴu5i^lkm/ٚX!|?tz۹p]2 ~//}o]=nIˍ#I~MH6L* *Uh=`w˟vZئfC +wZ&W=U%ow V m'O֚xStE/}HejP"+r83'5܉^+8mdIxmgTiv{VJDW!~0=SwБVgIz;דLR!l2z9h/b2P:ujeg!VOzlL䑚{>aPy^G;jIjyB]bKFkdz?2El$*f RMy1nRnm cIa(>֖[3w2(qɢ9BLB{l@ =vSAeHݜ_|)~}"M 4nU1m9SY NqQ2BBH$uM/ST0r$nt{.ǚn>RmoSWuZOMoiu'] tӃP|R13#^)CY5=.ӛ9GS |Ϲ|7km$g.RmH A%odɟ${.>Ŗ.m#).!J Hf+"JCćz#] 5MZ',ر& 5d!16EϹJ/"C7ثUsvcC&tQ)P`AoFKfrþxlg|TAS1Rh!H[$dFLO%a6P7qv k(.pKwhs hP0>o:/AY\!mX. H/MseRE;8D댉x8k֜=s?wdcYQ/;*`]Yi'm0f^5fL. RRh9OQ_sxQ KFF(sZ8_JQӂj5Zi1xGWby&ArL{<h %TS)Y 8l❜ "}iCԝ[Qʚx!XLH)edU=!8@hk"Bm= f&r3:ne BҊXJ/jxBy"x.F³|S 6gMdQ-0< 0۱fr谍G+.^Ě4&OWDEX_UHu@lJ&8m- g u(O-Ȇpf/[D gJ x|ȅ$ ](sQ5Eofye8&T2߻Oͧ>mtqӇmz#aZ$-%rN!fG.$Jl*'Hřu6 ōGe!SP&vsnz^]s#vu/꾜[P:~!c>tOS'^>pR&'/ )af޽Ű{k_ºA+m%Y[I[vc_wHygd; WBZ:>!/gyb i2b㫳dJ(IO̒ЖAykk>e5;:8Dq(⊄ * " DiЬw5z8>Dk=: yc$6:"DtE&b,&j YoGgv7kFQtVJ;RhUdžxrU:& tq"0M{oEW')`HHD[_ZIfV p5UZ5s/Ț׋[ْs;hp(!`ri 0R1lQA j5gglUfq/T}|›{[\Ud!qyBݽ;6O|s/ 0OG;زpg9Ap*]nj{IHj^i e@kii걕S* YrTMel2 d+}wGJ“u=vq:w(xfq֍`YW@h$hRd`8[lR0`ZSX!EC&d(Xt"hD&#OFtpĚfΡ 5g?IZ:ccG|kUIB K*"(zަ$ +>*x t!@z'87%|1VLm7xhF=7D?~{F?k%Uǐ;210 9tQȓ I L*zJRıԠ0!OAAQAj†AhTl I(5*5$D,rPh.,<=O:4r͑{9]eǎ{]|Yflp,BC)$-K. Ju[DmJx(񤔪&XY,JN^!ޠxe %z Z$R Z)u2*Р*Ci֜#Il:]M72n#Os%+fYcs)3<ꄌwF`\" Ɍ֙gDVl2^ Vā=Hأ>"=ʬE& 1*Dg %(a 'rT,4#(iC*CU ZHtH̼hmj`bt<(s._i񌭚{f<1N}8k/BsK?_~Y/bTF?tÍFAYs:ڳٴrb(0HܳgE;.E7w%ذ޼VV[z 2Y ؚ&)rMV x* "BsaYU F&Ž: ?7I><{ O軾.~nY{׏EkmwUJPOsݳA{wcEBŢWu^#wЂ^!_̋i(s%! RVђ .<ЌR @H2j6CXBN&  ϢfZ%AcM5sR!^Lk֜݋ig#[C;XiP\;+a76s _2\~_. 4qV8YRб"\-"xr6 6m2xWO'8A!gЈ$:P='o^B< ].YSxlȰ7L:Y Q2^o4Gyƾ4/jlҨ@<#^ 2׍pR&.K["B x腎 !xq-mq6'˼\>9ͩ܉Y^ˁ{ѢOOy%㔲9"9|1M/1hB?́- xZi 0uS{#V,]))%"y#, &0J$Y4Kk)} d``b ;zJlsvZ'GzoWgN\Y&2`J6F&M5 Cz1Οjm}$0sԭey&;?X|N`.WCMק}aw{5vߦW'E}ce,խ^D=gjߨBe sIff֭/$bPn4Ng C*B6:/.0*{]-}b:,oӴlCS<#CG;j/ݸ$7ܸ=n'ZyȭMWcҚUZ?;nR "*ͨt4[^=udқ2 ƣ7 $kֽj-%Ƹأ;%Cը\{ʌtox*QJbA7s\n9 !(m2/+<,fm3JJ8\+mgs(x}A33YJLR܂| ~r z~Ԗly% U{6qVnxQs[(DLs2dj%)KeCR)~0ǼNc\ODAm>_m&`]0Yj &UPj.mjO"ْ J'N>/>jH^L k-,Xj'j:=!5z; (%2)Q!&"3"22zJ ɧq'@TA  oFE%q kE+F÷/hX{_e,/ABO7Ñb ^6 [ƒ)\Õ}̯.? 2rA.rPw*Xp _ `{_y]R@Պ`qP{gc-_߶jWE$;II.5uĶUK@rNZ,H`IJY63jq35-[B lߘвN&W_{d~oh|{+5Rwb1oLJcs/j *R KXX>=Q0!S 0jvea*$Pg8z|~h:OҪI]-q{* ƭ(8{@5 9 T> XfK ?([) o9Vq$,s* XEd*(""JFc&=jSNʽbr0_/c'$cHdo{ kB Eɒ[ g) n b E{##˚xղ -ϧ6t`崢?z*p40{~+I**bFKpbi4%2kkJU;aP$OjOZ* 1Jã;S'hp͢t8P$Qk1VG ۏѽҩ n~Ϻ=O=; 53l$Ta$՚o YJ[C i(afy(q3{2'Lx׃$«iBjCG>^_ߵ[&;]t 2jX,Qd egI1=:뀳'fX,CIGQx$:$c+|9A)lCzMV^ wE%KfFJ~aaN $;RR_UrF;D`d(tS㊂4=:z&t$?BvS\aJ*3$1t$i/9hQ[FH(򥖜N)Ifx^GW/w[GCc}Cn W>\;%3zNm5=zHEdTA.2V{D>s1t6x)ud":J⭱ ZN)<ق:eH}{tz AEZc=1ZjWyUP;bƓHafNI?=JVĸA֧n[ ~uva,w\MhkvK;t\zwga,gKxEAFLH ƈ أH%R!IQn5Q4BZb)(o" JT y!D=snٛo9})&஋yx:_7Qo::/eW+u.ӛHU9CzTKJs_/z5%ǡA;n뒫]?7q7k}}ꁛtXH _2 DE.T,2*zhs5jL;kfἸ2^lD h{݁;*qaY)姱 U oU|ٸZCICmښWo\un[.<ዳG/[7 4'HtY8 EM}q2QHf"Jejciڇ<%dkrcAPj=_TynRM[!UÌ뒨\d2#PO(ҩ1Fm}y3+Vpds9wG37>L[Ά6XW7LRӾ-a:54}0 oĤ54 =2_:6nhNYr9"}fMrL,1َb$ÄZ̬TRi1*S,"c GZ'__fP G߀9rvF"'I=fD*9ei#%XP( QG,Sˢ^TN?|ֽo%E#R :?Ww݄@p΋Zfl 7 2rH8+*}p T5&dre,:uy?ˊUtWu#{\bͪ M/w?K㿘im$}ߜm^7)kS (J W *>|+ou^uSFg &d؀ GK_A(uK+@`MX;]qwx k.U񝿩YbEuq^, oiĖ$vI`W56ame˹bhPU5.:uim)heWjCX xנS·Β$og ( Fa_*cPքR`bafӃ C6!fWٱO?R뷫b;Ubf_>`8WG09yaAVa2L] =e3{0?M>YKv\֞tY%48Y/V++%w wp!φSn],O7 &wzx!$<-Qz\VtX/}7'$ݧMlC"D6ͺ[;j:R gL<#r;[; 5qK&ˋ4qo<%{oDUEU:oz a ) lMgajTC \  p,՘ʤ]]ZHj>?':R1fN MtWlL\ʸ[U( /|.s*%[sjL%قkݷ̱+!]5* A2䃡OD Z'* A߁2L5>~n̥?eU^{GdbwBZB ܢcمj]K.܋%5Kg72- p܌|^MgƩB u i8ޯ=Ěkr2ބɸ׏HŸhkyBp^gמaU%Bw*ާwټ0̪_.z0p:5d*o|,)sЧA??*R:JrEsݰ%/ُx^s w?7-;V9c`ա94ʩң@Ie.ʬN~!L7MbIL_@]d4P(LaN&TǕz]0KwyR'o]\y:QY>Gky!;%&T[:ǃS0%Ѣcg@ <("l{Sz [6(nġc4FhfA7L97L{f72:uJ/ VQRÁT;LLj9+j$?{`ٍ :nnس#{1Fz2nZfٖv/|nyPH%Ѡ޽M9[v7 ؖ`'TɂĔ#"oAs#Lo޶=͖_x>HrKEV9#0'"\9xjaP4 8 qrSt,92Ft 4HwƜ|ʟskP+1iivQ01zl4Q 00f̦r06Ytպiepѹbݝuoɿa,%io=Ҍ`]¬d1BԢ$JђdKk,)Q߭p2{5/«O'j21"Bd E" a?>j `<ۺd;)ƈ Se&#e!x*]b՘ ﵌FMFSf#'Cn+/iXx<¹tGD Zi8P.a|$g{FW!dtQ!`w5~ښȒN#bUSmɢ$˔-'`j5ͪ_U-M7Em?+!)>ȍ&~:=-al3Tp|jvt86T) ^ֹvlnC̄'YR"gp9JMrjGlU=wB/Kt>E̸iOfY.ŽmYQ OU:N+\7CO5;Ets6b| r6~YW2[2אdޒ1jf9 LJJVԑ'lu^\0ʸ8 m/=~؞:<6j$ÓXA} t/1~ذhST:#b&unV4c2:\?q߹[*[rfY6>ems{߶kl].F3x6[ֵ}8}4>f3?r-Oe󨐷3x6o\ :\Ƀ^ofۡ'_n1ۮMwo.F;dWC?i{%O=m֟#mt}l6g}ͱ:g?C|e,o4x︙}䛋^A9ts\Kg ΐ>86pF񀳽h^u&{EjRRN^ [s$4((OJp'݇w5%(r5cpe% Qȷh-1ڷ/rz6Qi)s2L .`x_p d UjH:ׇM3!v V@vNZ4l/:(+F7ŻT⯋??UU鬽ןj6_ JPwZqOE8YAϓ~~2׻_qŻO2%Nkk_l:^FlW=KYGj|Uc/ 울{rk/pmx;=ޤg{#%jf}GV=ɻ Pԅ-PBB|0zZW^|Tgʫ:NI(Mo脓&|#Sěz7 3R:)g:G8Nʳ9>ch`vnce@ma}C'gBsZԀ[;T̈hȎ}Sq*ӛ3PeI%Г99 'ߞTh;]ı)2㣌 fPZM';YK¯;--L)"zE+?O=BӨ%"c3c[3_NRҚOϜm)Kr(b]!KFRNOȜqܸT,,h}(2yz&:3:hpy-㰲x?zs{%#%@Te`Kx"%PA'Nq 5XI5:t ̈́RH#N ƃjQU(eh FΎ7Σq ()I)#q̓h.-A8H (K[ $7jER<ⅲU].,X-̲,vd=*:4SNxfZ%ka1$UUʽf2wddd D`'-"\#c*4%KIxPΘ\@f#2&kYϤgR]h+Sm8Ϋ^ bJ+1GӞDitTK4Ri%ZUw*Ha՞UH))+Bx/rfI:v0εbQ%ܙm!EN #Po4]/RbVqc孊9MS3PDٵ'ӽ4\>}`bVI}h.oڔӅ때.!zځ1־2`vq j]QmET!EM4Pn/@)̨~nq;zZ-X66nzH 5*KЉ·dJu@Q-ˡZEc˲}j>B갘@ٽoj w5}[CM ,%-!8Sl`S RX/uP QYnlc!RJQwUx\6p)N 'gu=%A{%W9TW|o"6o\{O) I#e>7%m3AE.ƒn ~= bHiaCqagN(Qb?-3;jqE<U8|xRL׮j\~箋>keUseۼ&pak2^.~\ ݴ _4<f_MrΙYlxvi>zD%Owґs)0R))WԸX|L!ҋü: Et<xU+!J4ǬLu%3#8Xh" L$i@:@JWJgQ7*kjvx ^W+RN~O2_)#r"/8{w&C/o߾>j&EK`Nr^Z!8Q)CNarWiŤT$sxǰ.W`Y>픋[ݚ ,W!ז \V2fpshxѭR0IC2hq͍.b xC|^qv>r:#ff; J& g_'s_{DqUX %KQI(Gqbfz5-{n=i!z5%!chP$7&tJD< lV;=DBvE_/;m:|[O;^Ҍ 7'u5QDYRʰ`]XG !Zr^i Lsd5"w&q 2蟼+{Aˆ6|z hX7#yuèÀk* 3 G-X%h8X,riz@*qIu\QQ'ձN FƩ|ܫ'<ڗZnyempvgo?_yًW߿__7˾yo^_xj_u H_v@h~Ckjho1F`hp]zk\S^3 c.5hi)Y[u$E(c{Q[.e,˖1H^^<`3r=HWPc*#ܲd b;5|V{tgvܣ|v&F2f0ؠH%d.S(2 BњVp˚I ԪCÄqc;^ARgg-}>{<rScR ZgMУѶKҖt߳ jf%uz#b^LPJu\6CPOv8&BׅM(jd3-Y lDj$E%MQ6G ,d_F@lϴe} t߃ufnl]M=¼x?^wPcmƘZ3>MP>C5.P<r칭`&P$4F1]lv$UTJ:)R%!TgUp8.eMQszȥfxCE I*=j (WBhF/%*,6CDD  @$5'$5 am"e&`ΘI.ǂc O%CH˘$Qﱱi)}(G"%+*F'/כ;f_\o[M\:e !9heHV$6jj79*uLBE$l`KH^).,vе8w^9C,_ewS+KSf+苩%ӏ\s.^⧹9D& R't*AP$H?)IO2JLr_ 1ZXCCCp>z+ Br dImq%|%H)QvJ+dD&Y˶)*bڦAXȡ^wLg!_*~r {Q/Z )UpE%+R,f]eQk\T$% Ow9ހO?03Eف̲&o[XKQD#їRt@Vnjjuj`'/dGXd2(A X"4!ҙR F$,6Z7c2;É${ pUDLӏ g)%л!"X#LL)eTtMuaj펚= E%>1-V6 "ZhJiKZaY$6]6F1o l Ksy,&.::ÐmuL[tO`Ma84[f~PդU$j" l@&=O]lJ!/q?]6SUc:i!{E4ᾈFk6>PrRPxuk|jVwƵ|VKhJ2LbF/< ɚ@V`9A*h7QL*YAOjNۺInRC[-W?r#zҰ70|!S%zןEn ?*mlo~ר]7f~ ݖTՖuE1EA{i=*gSGLʳlp~N - Umt˺4$TS>ɛe} ]8G^+ 2"h@6.f=X ҳ"UnCQ (y$A%IUDKu`ְa,{L7խN˩it 2Lh*XO&`#/Ń5~ʌ*ZfHQ|"Ig#LA%j jas՝6z Ljglpp27fڬ}:O:࿲c>`8_<ָ㙵lz|%wc~jc/.<򷿩& ^Sj_JyN&H6#dlb=2ZY1$krj^POZt%@X`.Z*f=0q0@KiJBa&5g 9cwJg3ؖy>{KܦU3^8BPwr1Idnh`0z?O_9V5lG ]ⴈ59}}ҘڂI!0Y)֨ 2Tc/d%tJ6uBz}M+H;؝s7ǎabb ;ڦcm\ކH!4ĪI1g%6ߢâeu%XKh:WS u.F[ȬJ`+:jļ&HVdYqڷWfB}s7?lEe#vx_GF]: ۖBlʢd'Q%R'n@zBXl-}Nvb|&Ȓl@bJlI+p:8 IuU٠[⭎'͜l]|/ޚ^L(SDTيXjҋ$ fe\p f&q$q+_ŽN ;nr,њldJ(OzC06q CcяfGs$-ϧAyuO$ s}Xnp \(Q'-B-hHb!;h:; ,vGh@WstnD$';MEgy6ٜeydNA/'PcxR$<-?].*{Mygh+,ZְM?;j#:Lo&W~nsқgH,fxLɌ?: C۽\f^. fu2KPƣ?-{sMo[o/=y-)7tj@EfM6iIvNVK۹vnR32E-r㝁&!`+fM6MB .wӐ@&ztS A&mv4v~IzIovC cYaM5b;Ps\*5n'f-OX ċ꼰~dA4 `{ې~c \>nЂ_h$JMsNdnLPYɂ\glGdzGW6D)f<_ت3z{g܋0[|Ȼk_)k[|e'BPf) Êy"wΘ9&dނ˹t|v-T5 /i]q_:$ qz,={˓A3Wjo-J:J?/$S8 ?X3`dd=S#5zeuيHzg%kNr?G?֓5Y2l6bn; RukrֈYQf`Fr s]zo,kp?vXSxiNIyfP eUBWyd=e`V I:#FɨLA$V3_L )̱o1',܏Ɣ߆znxiiFM\V ^=np"`u Xj+2ƪ+۵cA$G'y eHZ66 1fVE-~' iy$[li``yC hrGEdYT+/}v5A"bxir9Qkzhlv7}#𩈜FCVR}BrTںRm| bAFG֤ 9$x4ɨC p.5L̼E)+Y,&YkcRئB$}I1a]v~w`ddg\{65W ~^'H:~YC=ǣ&k ͹I([늁ꪷ^Ȓ̳CKcm\'S$~Z'OMʆb4Ye%VeD)js 3FU Uc'gL1ލ红w7s C Avy9<+fH57]?B0.P(SP|B C>EBȂA E U-(`GyRuk"A98bF`&|D`P*c*ʦ[C%ǀ:ʩll@^C2lƩ "3I_~N+%C*25cPHqHqƢ+X5jώPL@Z2 㠄bLcxkލr7Hmnc#eۊYUIB1 D%mNBL^D wgخ3qKղ=*jgƯZp 49ac[-k.Dj#ja&a%@P4UCБJ'. -VUz|阌$;`V7:`  $b{1 rZVȼ"a|RB8Lt~Xޢ>nh=ŋ "@:tP2Gm[Ef06mƢ*YȩՏ/XA}(yYQ"m ɜB]V0N+<]>?_f}Xk!O‚.&ޞ&X-B'G A]rmB^Z@mDluj /c{1B(k ) Zc22[{G:B&oC@E L3y20CjxrI'#6Q$D#+F̓N`8JDڢ tdK;l,$ frPpMڍ7FPٗ95*m@ Q'Xά=!YNf:zABs{1B|G |SdF)b*q˶ %.I%n% DEV;ͨ>rI"VŪR盉`ebBv"JF$| ID a8tĒc't\N&!K#؄r^VQSq1yTvիm~>^ܸX\\zbWs;BbC-&DuK`_~Wv*ԅt'Q2d/}ǻ{C>=Px@U81@0C9%8{4$'`D}$g؎H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H "XsDHfwB5kIfbD}$4D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D @>&m]Ꮗ?"*ٓ@("DHC ;"H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H "I Ä{T$C\ kI X=@_" @D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D}9$;[rr˼m5mח˛7vRr~{B 'b:kI ex{٢|?z}0L \W֥b\w8P9 !{ߘWkR4iykt8rѽ͈kGt5{?{=]t1ވϼwr:O&JYv}{]GN6 yjz|3=;oiݺ d҈ᡝ 9Ɂ(Qt] 61A&G lqe|Xٶd}e$[T¯] RA#4$p_r'L]]@iPYyV2m헆󏝨CYt7>[SEtJy**1߶p?Zriberէ_˳X쇭Ym⣉Y;!,LZZ _ԽQjM52Yeb6޿rxl n:._C㜅eI|؏Ah_?985:xy^:_}&jǒ31N"49Չ19+z$q4:t_`XM |jC7to샯~^C;gyp(\~wזUyMKX5LW&'v8(P_c? h}F+]9WFefr]2>9h=7~m6{w%ݔC.pu{f֏GAGrlU4 ϐG5Vخ6cC샐Om\vP|5} i%h=<b}y;͗ʾs9̦[70,WVH:Sܜ9.~MY_hھeXb\ נXsSovu~}| ˺$i^qC+B@yr{r 11RGv(tr[!oR}&Rm u(N&eXG96 T4k{\)54ƝHV/ych*E@r@ˍp GKR)S H&N z=Ƕt )fB`,l O#wEɟ[k_Cm|$^djwU,ӯݖZ bu"}rR긭>,'^wp}aIߦ{~<^{ mc*3vZ2岜ٻan3%cBbPmb֕͵!tZn:i~TVr;?,B.ys_E7,dK'6cCN{8-} n~aY5 5RH9?K}9Ңͷ'믷 &_.^o<,n0Iv:랕~߳HhWC]blJF ^I雯t:t1xe. L\>F[/Lm;ztzqK㔴Wn9tա:]uYlgzgwjU'm3?Q?a▓/M'izxqoӓ=߾8=yg'/EIH C&p` .g?ҍui\o{i=.옭t=M Gﻺ/hiY $ɏO2-pݳ٬l+J=r'0t}t)z{*>"܇q2`聃knD=jZ$I/NYXŵDAyPAԾ#BR|Qg/8lC^Ϟ>EL.ˋ?o`zq s.ϷgzI_!jg+X}ųE-nI I_!Q")JI%uU2U}q/|f]t2^sRrWކ.b-.@RI%z*&f8ir3ŷ b%5-_?< w65(oΛ=tgnZؑiy$tkrԡ&'7NXuR ~ճfH^L^yDlrˋ8t(ZTz{pKgfr[U3Yse<'t8&!)GƈLK}XUW} X:q-iA^z泺* :)H-XdBaSD aJsUFEHm"-Hv4K1Suo{C6rzo EamoV;?C?.ޭ'|| }; Ocќ!#b~_A2ӹ2~rp\\ͪrd[PivVjRj$$U0҇UAih-%4TfyX7k|Yw̤0e:۫Ƿ=\/00ϐ]ߴ>wZ~+#QT1^^5IU*Qp %sybJgS]ʔAE5|sz#/2@\$'ŀf)I!j5%m *H; s}_7QN\OzsE\Rr{-W<Ȇ&tq] )XЛ%3NV\6ǹ9x>$&+<Y+cgtt5l5ʧ55+[BfV&R3>t򞗌 xrXs.*$(J\jmkQJ=5S6Vu1A\qe&RxFxJ>%ijmSD6_lAaPvc,O/E;IȩkWsXs8Kp놰ՐPդHM^)k\`6/|r41]%;mwN[zG_NjդjWqNA|zu3ZNyJ=_iHZK)FT|@8-B(Jֹz[I1- D-r&vs2O.Jgh/}=hh3;B%}?ٻW|7|LN.Z¨n޽:Rw߽ /Gh[R ~q1k[Rm0[=@jF퀴^ iU{j&Ų-x|j_>@Y^dqLgYQ `9 ߻qGB/ UV%jҋ(("hYtTh6xN@ɹ^b rq*= $-T* dWEĚbPџ~f92S'M^S.oSN3Uw>Y+T[mC4E *ƨtBȟ oSESKj$,j@ pM 6~*e4ՠvRsZ.zq+jqiF7CX 3r.+Oc#Zl{1d1޵GzʓL.Sxq?<LkSjRU>)hwH&/dޅBcEx}NuՒhzWK)qo"nT%$>{JEU+<VE!Tkl9[vX-l&B>-ܬmqExO$bWCy^\?/._/;ؒvw!)|u6HNS_鐕/"eTG."7Yb+@d/UJRI2ʺlu8Ș gybܢ}ͤTm[m3X}NǨ&R^xs%*JJh:G)HFiMa5Ŝ2"CEgFhk l%H-GgdR;Yܺ\QUU'sR>:Iɩv5v]<;$Rҋf/Ҹ `pEǑ8q+.bo[IG_) `h&4d;ο#ʓ1n< [F?#|?wc5s,JPbJǴS bL}qs|0Ǔ}peOO\K0Gk9s-r|i x9p I:%5d!eEId?ex<|!rqC )8|K*YD/#l8u#_Po@YEmyz9M3'.w> `.3ڜ׳OXd.d %ico+4oc)L'PL at7.e A(]j9py\/f5F+Gy -:H2[/c^E]=+!g > A 1n;Pjernv3MszɾG1`a']_/KDoL֑Sc7<U+.#њSӬ=ds%ƶs f8͖Os2`͘K|݄8ȵi={3 ~xr8b[#l:=07|浥x;w}TF4׷zlI=myQ'ُ*PȻru~v&ຽ-s+Œٝ߯,|#8$2f%XY3R`>,I98ju/#d֘l .۠N='=^GYZ Ԑ )*lJMiPK5*a-GrTI+baOM5FHL*D(\u!T\`jmY$ʔ:l9j9cOlB w#g"sF/ ?ןzY+> +AL۞=5a]~U҃~g$k&OmQ4TTЕK&V<WosKo }s[T:*)T׊uw }%Ԣmx7ܿ'm}&]ߖBR,}zlP5:gz[/loƒji%%#K?A0e@xIã^/|fgw&ɒxY*Dm ^ \ue#oד1w( nü\~i1)F Rk:A })V!1A6vP&ʻBo$~ju7S(` j1lǗ 9\`Ĝ᨝WLz<d<07jKbeJ+_c5FY k}Q#]pH=Wl81DB:[:xlyyDV\9>7Wqr h+Zt&8@W>:w(z_Q/e`XNgˋ X뿰mt3^gZI~a|8nWQ݈P\t]FL hy|y*ؐev⏷uwa=֠_@cHu">NWGK ,fd:/-Zo˲=Th F+γ^*P,5I $I0R3XJk̮@f >uZ9&Z CAT^=&=ާih(yj&?extr͚X~Qs=H'EC9]&(tȢQd ,E܃VGt0IMEQĮٔv L>NpcY00p?4h\jʎl})6rR kB&<1p) 9sg,KS3E!H5L#^qz A96"Fm:4$J^D+NJyi 1.iCscCYB ?FvK2ŋMYNQJ1}{ճTii|Ȁ`(Dr`F(eK zNଓDvPEނ.Lۋr5P$%\pc1NBGFd.61u PHjj|NDQcTVFsLL' ΙRqeMdߴS:x|Uo)Wt~qvXg(Vw!}lI^CigK8-.,@6jҺz_px&k5l-i״vZ8f4(hݏau~}6ai?_; _X7uZ={EE4py:=;;ytVЄ“Nn>9p1eSp 4$Ѹa)U\bF&*CMbXh/,kY2IEC`X Z}hL\Mxw,uҲ4(鉿zA3sxOn] r! @XXٰEo<폠J.zɁ $R G‚\k. w!3&,_8';u^_MGGOg5ax3#=u<2zPY{Zj+N^PP-mPTwqgq]VtL㺓szִZ9ibǸpu N4e(D٧}3eFuloJ>ٰ7㫭E[TJwa^2{ ۵^X4Is5w[]|E=79D'1;BNf)lgLQ : .~My9h hsMg2a~# |^]tcU]j֘ m̷4].0D}&v*Z,>s͖ky۹7kȥ%稤%4Zee( sĹH#`ݩ3Ҕf\G x嘒Lkd*{;#@i* C+*\ :{ QmrH~ϷPjE߷9Nn<p0<@?X gl U7K_zXIUGg{[MFL*3ݻתQQ59Iaz&~Mp7oܺכwq՛_J޼޼}~O mq}6: oA1ݮ^x@(&SF?.,tb|Oišmx'ݤvp#= əQP՟ĭϲ;׎5W$]1! E61M/Uk1!I`x@?a.[t"٤eZQTM>.3`^@<-.ۂ#C*~1A!1c mdVQ|C)_Oi1}.зW (X)A?vmCx18-ξS~B/sKʭNVglXeRTB$t߂@ I'd>lY ݽCcu HeB͌p~~{itEXk%JR23ZoA(k&&QJj6p 03#4(q4U",0:~6(VM-^fO:[ 5럦>KRgas z*MIC!1/Y J3̅2K\qULYVb6I( S~lȬ&{Hgԣ[T-W cԥQ/c$cAQ{.Ut4G6LlaI(X$9J2s&E2]/c;$cc ^bUwWc%]z> mN+g:2ϫN}Hԋv]v)eڜFVdj k2r5;ĤaUgm(2rX&-_asRƫR!D&,kdd#(PLHi{D$lWl]=Kvuq߃UȶAt˰\'!e- q妈Бѵ[/Hh@g@1#ck @TlC풴tG轴u%=ˮL?t.jfB::$oU@Nܔ( TcͮVmfr1NZn8UǪb^0/)MR3{_&nNxyg)B)1mϧd$N3B?b[лW`ǣBee>-BcjQ+ SB-"v!i<[4@6y󥼬eyIq"F~h?w lcn?[ 'ަmq2b# Pl> mev֭wz?~~pyv﯄i\Q}?v{g_̟ym6%>9շ`~;0>( QdN9y'mB8YU%(m548;ݴf{O>yo>9Cc=ʹ6&B޶{2zmoo*[DNԹ0>XvhC΃O4ELϼJ y9Z}K6Pi GyJ +(-,ӨAFEpkIMe il)ڴO:ٝY$©)M yڸMLJ0.UJ@`\ ›kߩ(g,#89jUZq:>Nz%o8ˢ8DT>Ϋl$- )As'8TM'^%ގwY4Y͒yF`A*[&QF'I9S*툐lWD]7c1j*ƷB"dI-)J1 7LEd\`0.ӂ|Sh.`yzgK0ͱXG5 :le밿MЫ,1 KyQk̛4(!'zo^"i6ыlYކ)%4*x1 lmYHoL?'lMY<@Wb#BZ^ *j_M*OB7>mA]@YNx, [qmk ɩ}&=:ٽWYױ%x %KY(T2X(c(*Rې ,z'+a:' W]vM)dt9;$s!zCe g mb5qv"S9NiY*ox%QNkKO ԗ ǵS_gl JeZǗZmW)m'q@X99KC_`IYדIQHf` ifjMG@פhhAK]ɻr>IX'?q> ӂ!]G i j[sXx2+OܝwYC?uTpŘr Y! [ d!BNOGsr[#5 V;B6:/.3\pANgfkp7uqXWs:WEd8qh_Z\ڱzEkd]2fsץOmtNc.*ԺKfzs4vK6SmͮCjPnnG=6p=iY&W7.:y;l??=QoHUTMwM[jod_xEgE\mip;mBsys!ԝ.L =yõ/ *Q.B&FIr.ĄV߉m8>.F=(v4Dt#s Et)V"r1Ar` 4#;qB h.*a*e` YR F%Rp2qDBWމ&Ύ'bs9~*b?ZO.'s!l3?ͮ0>bg6wps"xz!33E bV8Ζ֤198X2KNoxxHK:HDr6pk@F#V&v ^ހ=Kѹj7EcY(>#,U,qԞtd$5_]#zE !566u1&hZY}GG=#QD5JqP Ȓ؜J A 1E&% 1'i{Q]R(}HȲ`>YeE;b~ k,Y>9\Kؓ=|!`i)=Gsx4(Ҿ9OFT*xUnTd Bs(lr= }S+֫J_gH!!}HFR2:J1hkc%*q2>=^EfߎGN9? I?Fsoӵ|, )j8˝\JPJCImaђ ăH LH鍷OGkm%Axhks+ 7=X|$_NMҼI#}V# -弼isQAD2 &2Pj^ETc;بT80L)+Ų0A`yPȔR.fK[ ]JQV"'E:&gZC8%ͼT놳-٫ˍ8_2<]eg\ЧsNf xĒFiQ {nSw_=[H٢a,ؠ"& B fP4)x&59,\TAx x\GK$UP!2gJ*,vZls8Ӌ3 x^~K`m0|Nvv:2'>>.@LW_ @Em=#YO@+D^݈FCF('?qq@듈 ) '!e,9Es {$0tF!h\ґ+T@%FX >!!t1!=ˀ/cx_vi6O]Z'msxˊ-+܋_>>cAo"ENb31=S>e2gmȲT(cYdPb\{;p "H x28@xܹ$ ocAuyy1yqDs>my-!xCinͩ;%*~7x߫u<"L  /y%qg<-1G8:jygL(q< {׳yiC ' DTE.X#䣴+' Se7zS}qxV  tB7 tFiϙ^o o8 Dø٫({fѯ1b?j Uvdw\T]Jx/szmY< }d0\b Fm}(Pn>U'Ekj'Q+_[$O{o3H͵1YO:7 sAWy}?]Ҁw.ܿcq],S#Y\R(202q+ܝKaݨffۗŇ.ZQR!|5b~xQaW.pSgߵ{QQ2LE:IZ`,5/;h&:T}Ƭ3˚݄>E14]\N➠&d铳Q;s8Whm'bO:ř9Hׅ7uȊ2dRṶsf_.V/{júJ_L }jǙr5ZXYҟ8Cswf@nC(TnwϤeLV.)l1Z|?,;K@J0Q>C\@ytקwgp_FcZ~u,Z_.;\_S^ d?Ň梙ӛՔ/?.HqX8͋ U.·7Ki׫gZ820}Qov@;F),(_ME -UK{K?An1FϏ:ؿƯxoܥ c,\*wE5ÇIķ?/};;풒h=WG|y߼udVhS prtKQX)4ߨ.reU-wU65p4NjSyu˵wx_&hp$OJ~j/3l6qYd\>y Ox0:g"pcmӭ\_wo<,⬈D%OѠs@[uh*% 딏)D*,|U|{-p#Y~KxU+!J4Y, @=rT1\IZ"lj@EE/sOu~ך5X/ ˋx[`ًASds½_qƎ%C)g츝ơx\~ڎL{ X ;xibYUjK9-(Дls_?# }(!$Je؇yy:{;ozX(ٷsY\/<|~U\vwmq09RA0LJa Oa]R {V6xJy9h 2P#dB[L'2a0\n,E\6 !D.ysIӁ;6iv#*9S?饷elZ):<$4((OJ'XmQM@'\! -80JRQk" RіD@2_' d*PHx$?U=L }z;jV:tޡóqbW7g;T4?^U^WAz[Q…^fXu9zx㠢^h)ſtIr3+ W uIcuNj͌Py<]~#{ 4xu4Ⱥv׸#qtLGs|H|XusL'_fH:>Wf̍S/4_z9V~rMe OSkRƷ'awʬbё81#7Jw%#%@TeŀmP(_2*)N+ƶ, D:υfB)jcqM ƃjQ!R:)ee5r6t :&lY_$PF$1OB"^P4@In "+e|CB=YB׹P& jAdeT')U1Q?:%b iM`TYgĽ_Tn-W c/#J"0ςr Α1%Pi,%C9c^s4se"c{$ch-*e[MߪXS/uV>㴢W=y"BC?}hEkoJFGU; F+Ֆ)[q.J`aU; )RrX%-/Ⰺ 1%eE2\,i#h; \+QIZ[/u^vR%ߜ2Z۴1akجE܀K5FVJ1 VE9lhbx T1Q"{j'dM,-a۠͂F p,+l\Cl1Z|'YeȺO?6>75LXE>Yqpsr'kPs~чGFԩ;ҢR=JE T4.pϖi0gj hƠFU'Q.1X9/zc-iyIK@rh.1j? bVIAy\D0!"( [h0]MKt׉7ϲУMA&(p?KJKd2BK P.spYEH|tpl^v6)%x4SLSV9:hQ"KZ =TG(x{"\K(ja݄CxGsL= K*5VAnHNq ksZy)$ A>g hP ᅱ,X: Wш:ZTV>{(,Em,Ԯ m%EF::XqdK|DU@@I\$%Zxzpc[D/p-;o'Huk' `(2Abu//4^UM2xsiqʍ0epV햳aU%aB|yq\Qk6(9~eֆO-I ';upG0Hh~~<[CO=8dhaA `\|51ƠOן_xݜzÖٞ9*P1HN& (vɏXJ3Jh SZʣgHl`v9"$sQ0uPV[iYd^frA#-ZP\8bMg8omR,R)ϥObےs`I? H3n>;8/k8fR +׫d (QR@1QXQjFENрqF0m^B0 mU!~EZoqNK]~9ւoUt_pjGnn)خd 9]9kձsMH wxʳ3L>9\'wPՒPD;y 5 :h s%y{2#ͣvsj緯cx~y "EY)#IF8`r("F9bxq<2Wr'$3< B挬~Kau%۫E;q,˷D O_ُ{?OJG:* @q82Ioow۸:V:z üզ>ِ`= ***C_]%ܤ wxWwxWwxWwxWwxWwxWwxWwxWwxWwxWwxW !rB(******bg*FGiNcpݍ(6j!22&Z[!Tm (@y}Ds["(F-CAhH$|vKM0B>z#,e MA`A/ .GYM, ɄeM (xL ehuĬD15Ξ|v|`y';E(2$N0m0&Qs̘T@@B(cZ1ѿrI=֋E2o^+Qtc8Hjh1AOz'7[ ]5 ɇ% b @L`2Q'9hmQE`7EJx9Zd-s",b&#()(eV5C Ģnkg5P7z~,׌De>Ә~+8 %V)@Ѧ$U%bt12M:kR%J9.ю=nXG(PWzR掸I7VM>BVHTClF& w6YoJ:^s[fc:͋[Ep_Do:Rj|p(K64L;ܫaSo,0v59|nmf k9d`^oaHZ,c9)I`L dbT"L:J2LH]l{@H* =9˷e82>Q=F_dv0'B=J4C}//UǾ)1nۀwo1#Յ w-\M)@gmS B̖Hm8h{\.%_i!Wc~| iyʳ@r|uz>w%i fJj!Iy Е\m4?f8.A?q(*XV䕶!#&zǢZ ļr9.K9RC^"f Em$J ^Fm4+6Ξ;E1zT[:lNϹ&a*FEbYtRvѪxX۔Q3rP!IP>};>x} 0O.gCt*]JĔHk]iO%2D2 ֘[9U*{!P1*6d0L|u;kw\xhF;=7x?c\ᅽw1)U|2Sl:i˸9Tg\W<7ѺRH$oҲ1[Y``=k>Ik9#7ǟws +ggy+L=#zruy: )4 AtLC^Ή xdK5& bz}oyl~e0h,)dt$=bLS9T:Qﲩ~G7Jy}zl_qYX/fKaVx^3BeRIfaGVQa ('RBwZ69?Tƶ u?ں[=/dD7mNs5e-ݵzmtUJbMtG맻wy{:Aw zG\Otߔ([ /Ud? Xe0,m [1]ɮ*=yJ'?"4ıv\TQJ-&`c%Q'#K4l !4I:媗9QK%Q%cP9B 梤)  @@,] l8{br6\|r\ίWx<~*i4ixJ]Z-l+z+XHC)ŐPXp*9[Y(;bROTLKě4r-6DqT`@Qhl8OsFe[ΛgqYXA|#+%Es2M'Ԍc^F[h S`Sw>K+8hx$ڣ>"QI˂(M09jbsf1E,AIc+GCa*YÞt CNhm`b5 wQ-kPZGؓ%twOBԇpw6l?Rb4n݄ ,`\t4nv¤fI.]藼\4Fa4_GI~nrgMOfUB2)H-fy)^i^ 1zN,,ctD,w9P^QMTB {*sͿ_^T;c3|^nw/˯z>Phz_׉G+rIl2_{MRb]=(ori,(zP/gkjޟ]_O{px"&[~Bk?~|gѦ06YyC>Y-:BH;r0gmsrl>\x+us +T ʐYp4p.뒱&D! E=󰮑~H /VB鐉?rϊ UN1aJJm(Y{IiJA[c0I%d$->a osHE%^HΪ"Jm(=P?6.Iy9 ַ}ՖaݽRilwk6qɷ1OJՋA{rz㟱E2v|NsP]T:#XNkV57>_XX!EIYdNɔZ:Z҂a`#0Jf }[ai9J m."Tm,ZڢV`tPD5D J@j|l8O_LM[&:>I{{#}ⵯ~=aѤcYHz@Աc3|YJÔeR2&`mm¼+A|W-6yBA#}IEWv7o4ʿ~s4z~3˜%:OYmTA.(w4$"AyZ}mƾVcIgʕd)( :o@ˌ^v3>ʤ;P B:0)x7gʼ#ˎ!s^w{F?J|=t/i!8@7>F͡O/Y8' :C5q.F:d9CA+-y;C!H#` 06`PMahs %I&,C[/|֊ .,cMߋ!ڒ0k m * @ڪRY&Q(P ұq A"%Ba o b I*c2. O oU)/QjZ>oM|&1cpD9^n;+D(dͅtpCdwTC gi=[E y>Q6iDtlIh V 2JMV[o0N p=иhAy$8NXr{w1 A՛w#@(gP{)<ob]moG+7_t0q:F>d_e(ǡ, FUg()Z`Þꧪ%AR EN ],Ic= a-V ]D um'9nNdGx %it 9^;Wڒh !NGLNJ%[ Fdڡ(7* w\H(aBg5~5IӖ'5;+_/GՋq-L0R"E#B"[G개:%E4E/U}|-飶ޛ_?Lkڲ,W\ǿ ΫYE}Ç?; ؈P"d:R)KEc Kխ&ȍۧn|ϳx?WgfGD OfYǠ${iwmw.W4!i[u<F9}[OpӦQl<|էcL9L OOom{8RV{s_Yp9v:@y_ eƩz?fLFK- yY7 s %su +S_^wQEZdz]oYb5jԜ^.v) *h.\kZ4Mn5.֚T;߶ZNQCU<)-amPm^$y*&!VFmMLh4~ 9{6:~Ug":֦<T혦=lŠ;Nޡٰmv-=441In/mXd5fy;:6~3mzL`.``.N=_ TG34Ӗ9"F"'I=fD43h`6`KAxLe׉KIYiUx[iyo\sޣ דnd\;PӾ }ӟs' .Pe7:9ߛzɢ:,%Q*z)0.wX5Ϝ.ڷS\)wMt낏VR& 1A}0yTQ͍QmQ'(߮Iqvsڽ#ff fL䝛0Kqe%YZ68)=LK*O;nC, #?5JesێTFILUA8{N MW4Ax4ŀQeb$}GX#PY33:nC'fz ^}Exqŝ(? Kiu1,̕ pC2`!hOE[qH*X1I>/.f脢E;-pѫ{iyЛ] k5fB\RUGC'xx>(j/LQyMQٻ}@0쌾c}x?BM6G!Sx!;ND/a-NRL½1𞩪pfj Y/ t%ʴ "S5ME9;p 9جkkxRbG:x3(d^$~}~k=IIk.ޚj9ûbcև, =9h>( %Z#cp⊹aIX>atX|Ѥmv Wsd{m wSq)mPqݑv}C#M 92`Oc`I(a࿨*H1 bF`mUHXG"\JlB OA' ¨U1kQ>@R5c09o.{B,AIR%yYȂq$rH{gWY 2ܟZ҆՚݊_HUk̺3#C*}5,N6gB K-!ҔD ։yvY}٬ڒf҂QSΝ㎓YngJ$`c-Fꨔl|XN H:S+R*&y >7; n5rFR 5&= ܫxb6 oZ[Pw6誳B;7I[[J,s2JJO'XjT ʔ} 䴧%r,čd] ]ݖC!fH ԦU>9HysQ b#V3a??}]̜`w:.g.l](IOF Շ?)np0e*hWZn5xNKEzb{ g{JY Fm9vSaz!٤^CؔhtlzMyZڒ-Jm|,ST#,dWVzyh>Q] QyYp,KRqQ!ù1b-(RDEs[F؉FHܰqV Y1շsp™1)b d%0lX1tF}d4s:2M A$uw{oIҾ"l楆VSr&}\XgdkvW-_jZ[z)=+/E#B"[GT6Ց() -zC4޺lڒ>jye[ödBi渖y:V?wM͡Ňw,SD@{tRƗz [MFӑ OWgV |i۝J7~WvbxN[ϋ_pa/7mBWg囗>ceM| xmxb[Sb2 -Bux%Vhm/ˌS76֩9k Zyn0' @R]uvx칆9o:]bmcM;i1@ӿ`gO(רSsz9إ(ZdGExCXZNQCUtS)u5PY b: /xpIFR0AmfӉFJ$Qč.TVz(ԭxRUxc~ד^oݎq몽=8[d(tHwmF4݇#"Y Xx':@,byYcKQ#XjՒc }}eԒEI-1R&9U+~.xfcd?^MǟKLY&־]/W5Ϝ].kzycAī^Jxlzq jrdeLjȅA1ykb+)HlI^8=#Y+un#y& jvJKHAL؀:VCj}R>91]^BWF]J\TZ>Pu- psUe3YYsU!]u^[ӱ94'spc\/-زzLA:ќ]H,9TYip-WeDWf+x{ 5'ѭ;iX)́|˚(;MZESm*SnTMZ vd#@g*Q8vUf CDމt f%x$V*lH>4;r U$M5'qyO3`ir>mέ 8~j4rE˳'ӛ{M87go n ^_YwDm}oOJmƑliGdڢsv:+YYt9 !ANtGntg_< '7 rM LG.6٥Z0Y6RrTyS;V&oktKe @@.hTJX1dX\uXC4q_˾NJtr%Os%߷U ж;;}/ˆk݂xnUcγ _a6Y8,t:YƯN>gO4&o /:'P9ĺ;NE'~=lA K_F#=Nl=O7蔿[Rd3Xjƪְu&L%.1%wo5"i&kK%2jrJ^k)ObmfP egp闿clJm 4rߺfi#ԭвS2.$/'oMD:#1.̯5SH)U+HP!6* m:_ cFfTSɨFuكֈsoդ$#KJ1:XS-dsB1!uc3(nT[(B9bu9)[,qTB،.@cN@#ZA(0/ڦ3t]y%8a=35)&Ws6@N7B'!e/Z_zl~[7쮸޶WzEC.Pcɡ:'j‡P(:1+'{h{2o/-xQs..Z&pHL)(1hń C?ثt^͆.tH֗úLy z1x|V/ss($`G%0ym1_3f" 'L4w^[hg\[]Hb!,ؗٱ6E_l6pqB5jUϢ"Ȭ[;2M02'M*1{r.1u=l2a;zEl 詨4*rkf|tJRB6NGPN?hQm}0, Τ`A[rґj V$pjUF j峣'iy.g|Dd[sor B ,KWPu&_B]@z(k6EasQ~w%ZtQ= ç]юnvF'-9 =a~ʰ<5 HM)lz(0L^…"C|6Vm72ouzT-L: #otgVC\jHEFdm*%k?6:M]_)\Phl U6h VSn0C2ŸoWEe6v5d9*M֙9nΊm:88mw*O6γwil tm}6ӼwkoWټFǼ\yop2[J)EZts]!ςd\rQ`B&JkL3M-xMkst߭QGRSRd,T*h#3eTQP-ͣ8 jm1V1R-cmandlk ->-\Ϩ]yC"udԝ}峣替8aGG'ggb+(HʃOT=T؅`ak2PACYE;bAE bd4 ga&_) ˜CKɫʞ:[nӹbiD1wjwmlhG{=ƈڲ@)/*DBߒ \I1޷ZcNd5 CdQXth$ΰ8rk{b*ն@3"wp_m~Uؽ]cE|QYf PrUGvPX]A9N!jm N}@J9"PP'fgRI;-r&mCIwUnӹGU,:] K%_g7).vэvq Uh*T-;W> dXAێ;wR.Lvb/{e #u Yޟfg|<_#'/o"bx 1=-DW9&K_}{k fa8(?ѐ2}5 C) ضhWs`t>2M";\K9RB b !Ek+ bC-cIL%$ln&yMzf Ne?iÕ㻒ܲm[PѠޕï: <ud9ie`ˈ:!WrA:ɔe1Dnq86J:c6=[o]liNޥԍ#B_5^f_|޵wt.~#T 0D! PE3 @uvQ%z{ +y5s!)wPtCw+(J v7o@ PoxvrIlGr_t}߁  URLfPU~&U0jSdgˀ@5FTaMcBDՋ"W>x!Č<4LZml"UЫChS#_7mˀ1 ~lTݪ5Hۋ{IJ&)%+^°eݷ=zVj<Ͱ, $_MF3g?1Bω!s`v[3/O ݟnӜ?t$h{= W>RF>=(ޞ,.^mñsL"Zrt\H$eAr;k{ע{gky Ɓ9ʺHQb+anRV pd*# 5o OȔҗ RRH&'޶U頧d`2g?\sb\YtqvfBuyrbh>%o{vjϳ]ҟxM$1GRRGs3|Xˍ-u NSvJud;%RHS49{=Vf$9XVp)Rq,ӊ![CbSrӥm)F*7׼[GZ N YAC0{e+|vqh\7%Ł=Y_R\7!^ 1um˞UD I",&'Ib3ߕGj%S.c.+B(:xfyRW-s4{zCfn +cȶN~IEml:&{؀>X}Zi&.۞7מL>.7.^\jYkW}'-'2d6dmEo_WDĊ\-c{%E706DEy fyvVI={*wyq˩vh1lh; O37~Ki?G_Cցo絑;M[4՝&kv?=GMOkSlD}ݏGuN{zu}wf@?vW~?i8wly뫻{|\=|σT[oQ][=o`Rw7yҥٟ.9}6Ɵw |pF艃PW2uS|YQ'64Ǟ<czF,zLbMMVULKɆK&G޺&Y%#Ϝ]H%%화#ST-}%V]Фl*XIu> XE'n4sݲPSejM>L)wmkgK1i.^T460GdtYʬq9Hl RES%>DN_9 [T{fʳ8n uO_GK9nK)9++)Kʍ?#Oˉ@)*Lɚ1ԣ#ϐ5&w :AKCNzd/u+%6k\tsP EU]v {ؐ6:"w{M yP#{lGnyq%֩lb].̊nY{ ٗk,ZMk ߗ}'"4skU@վ 7\PY'?Ufpn3@X-.AOXìUl:Df. }c50/^/"_|00 Sfǵ/U֛. ZVU|^̃k(4MíZl5\~yig5䏶%2R3ր24ֱ :!?~]xHk/{zE;nOl֒ghWh +~:xmF7{t2boQ!=Bguגn~⼕=Y+,K3ڶ,j[=QtgTG\[bnf{T@>O k~Y`̔{'&FaKDEzZ+k*IY-$3m%[SҲpDjsۧRתM[֮Py9/]l[ӝLs6U5JŐ@`79B4Ze絥di5I Mƪ}N.I_ CrU(#,۔i'qgڄے%U'|FI?T(p02Rl쯷)Vk<1c߁֢jA]gggӶ/j 礴u))0YLqL5$\Kc.>5f4cUJx38f5lF'VFGX*uLLk}9Y@e;_鐔m5[N'46IexPQɔ* #y zUF\k,l 4 \B{ {jI9D{Y/科" |dh` 7@ɐmM`[4~:'1X EK'JdUKuH(H%eFkնS7j)*!;:b(x8mH˶XaՐRP+59K}2IeSv 9L"zԼȚY{E/uHJTap1(jKȍQ1K‹yX fU!(dGoEBW KN"o L!Dtȗ WhZ-m3cK.!vVa።  <(EJ%rXOnZ ^o)mfD="]]@>Ih|Tx K8 >n:@E+(Gƹ͇ɐ<"Y7lq_4 k AH!22!*FA>8M>)*2]6h\h=G Arѡ^F=حVH [, +dyPunT& :U(U Uĝm df&' O6V}{)~_]"<$󉥯[%}6Fx4= R.bTmD%\ۭ||EPFt5&C0VY$[.F(ZN;f0]xl;d,53`b V@МHEcnXd$@fr$)Xd9DP?zD-=;䍅""kT+YBy(?* >*|LpY$,)b&6`k.'`0VQkYA5Vn(3xj m3Ui[T[XP@J6LP_5mkٲe&c6mMƺ6ڃ{W3͋|\.alZ06eM5 D8Tqu3(ip(46&7F)xv ,Z6YT#yHY5:p4vf cҌ`ӈn rT16BD;=jP"q, X&RB␐ @OOW\-a751l[6@ ?f_a+Br;X<* :(Pb!A;g4SP"0RS*d LYz@U4hvأtnF$řЎA48v(pc֌MZ4Xf=P5ɳmOdIpUfm} pmvWYvm>.IwoV-K:' oڕdY%yUKElr|g#Ut}<.CɴaiMd&BVt %xڠAL,R0 0e6Gte:)Aqt/N}.}[]X%E2kurKIE0;:@r&!x$tUXI.x 41yD  N]( 3T&+B8(X\7Nte؅5x6R/X 67+QM}y59̽<Z+s(pl{ ױi +<zؒ^߂jc"`xH (Cw$O+Of$Hn@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H}$H (<"( hH (OA)As$.4CI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BIϗc" s9J@PJ. I NyD0$BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $dIe#"R4GCJ},$P*&NEH}$`8" $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $C g?HHd~PT:>.7wo@}g~x=,Wm@Fד^?*;n0 s N8:2 d`>KI_<VBϞ@@z:m/s ?:.{մyGHUB$8(ϒϦQA\XS0yG0 3![L0fσWA_]Gf!)cds dL 9`n܂M[^wQ ?7qa?ih/NɵtBL<\i?Pd3}d<ꗏ)] ;lX21G**`doH4GelN6 1jd$`'5O̤N6(\HmتG3{lU1`;a6T&Al[ɦX,mņbz^V=KڬGinR4٬K(Ύ+twwԳҴ_͚ϜrFsw.LßwWAɕ١#pevT@]+hp99fN;]zC]ؾ;) vqu<2h,*&A[FUFc,Lv,Zdh|jZn/c\̡6[u؝Gl׳ 1.! l·Cޫj.Vͻ.cQ󦤌;AHi` 5VGBT .2Da(wJ-yh΃_?{pV&?nz컱.~GZKYAjòIڵXm/f[KN{g兩|;%% b`t>(cY7B-.ܗٜ}\6yG)|cS?naIMh|] vB)X/V} oonMWܜdPW߼o>wy{VZ1v5׃zRj,c*KqF`RYmwwm]J~f1֦.۳\+{󷋸`;u?Qn. @߀җ^~vihZ^0Q=Ӿ}p9 ?? e{7r &^sT0[|EKm~c{. *hS6~7__.UN.fS0]Ol$ikwۣ*TyMV2Of2n_޻Ė5ڶS ?-PjS' QgMшPˆ1I t&&MMlM4VN/Cgloc>֪ 4FUGtJ(_kenvrǼj^>Tjܑ҆ QGvq%†tڈ48#|w~A]].Ҹ},´\,KrDz]2,s#]>յyَ~ϭhҦkٻue,{A2j>W&}m?gQ_\YueQF"z HU=˂yC%R*^ԨH)"ZBJVsGdcT @1 Su.&\:LȸE:&@RZYD٭b*ΊbIn*yK_<wt{!VqUa49K:T+v&WK;Di꼒"(&O7g8\T=kUjOMpe=S+dJ/PR[%'t&IHk5O''ܟ/}C8}bq^N8i)itO/ ihpň2fp Tqw6pЯM5cvS !Cj+kϮ. $LE!;Pi7f7j/W E5_-퐴Ui+Թ9|^7 -zbi3uUg]$o?.3 TYB\vU牵o鯳,@y:-t6U/'~4}s̏Y]bpfcRW'$O ח7wՓÛNGn6_g&B+?~҄e_&~w 7eɇ eoo~kF$#!S`z.~n @,eG)ūz>?mzG 4"ӷuYHu?y:GOR%oWu)?eI hqB Ǔ3!B;b^$5cv4kjIYs(ԺYz.ӋJ^jOF뢶Z/َmuj/ϧMO Ejנ=)Upsgz15Bsd;;S-dY2Rp[e)+hPR3xURH ֝[MOWPXqP()CZ^_8E!Yhǫt!sE 2=d\F% WHIPPjPtW|3)7\VgX~5lƯS:[]dy| @oӤ>{~s6#H'v0_Ux(0׿},{)^8h}1_}-Ky=I=y*7> G~X 7A /O0MwNi~]q:1q^Ao;Is; oIo5>xw/RdiO" FWi}So;sL#=B&fjpRezd/2ޜԹy2!00Bidp)]d38C| u,Y=37U)1|YQ}tsd C7-[d !͛,56 ŌԲcFGR!*lD -$!Г99%(*&OlU+pRM:}7ڟ 2cCuҥ/%;YYpY[{]CyzpBf~x˞ CJhtrGtH*33"` }̢, 8:P& Y۶uǣ~br_հNj}.*_JrF5DldvixTo9 (8+n#/&/q"îf] scOrGUV_5Y8'/G|:[]Np͉VӺދ^yJʣǩ@Y]Q_f#v ZoB+Eᢵp  !||)q(!iy!bL`m .XdIuD)8kH\1RFT0 v)km:fr|L;{my-W|Q"!8B P5-!:4ނHBj[g,jD }= 9)gI9gtIr=*`zWHR`A MV5ȿ }<.2Oifycfa׈"(kQ[i` R4-Ⱥ6Fo h~G^B} QפS١2& B*XNFQ%1F#5R(HmiZ[$C@1L>qU>'A$b(.pڛ$HB >_ٯ^G|9&ͮn,t|Vs5Og~ R- Ձ u뿿чIa JWy7[u4'SV+4d&T_Rc UH%Z^{/6iR͋Z承-.6-O??ӛ߾z޾g,~+:8Ć6$<> ƭn[Z{ǭ|Ƶ|- >b{v֖ M~6Rx2ODh?S8l-r]pߢ5vuKI!PM4:a5~p+aA8EA\ >'ڸ@JTKH!CsH^bCb]GeD#H3QT:3"u2P4Ѻlcی{:!^6B[tˇˇO&aQeVBd&^ I>!{(2o~ V59SZ&Ϳ̿e *DPĜF{FYZchE݂J}>RvIt'gٌOvڳz_0csZ%)U׷/:T+La!A0QNS$UQ YJ%FM`#sR7.(R 8VZ4 +֚9wkvX]gV.ԍuu5E+2]wgm?ӲW_C\c+Cx')/N"*FrP\1i%JmAqBamgvj 2Tg/d%tJ6uB v#{H$*"9Xc7#ng9݌;vڦ6- asy./:[+1D I:GZ`ôBH`C#5DO#:Bb.& GO.9wީ_V8،?F8joB# c.IjPIE\mK!vaeBQ4&Qd'lK!ɀX:X9&dJv }JI+p:8 IԺ9wkďį:^wKR]chG88ް"S0U"9 +Ʌ2.B:"VH>qjzq(w E8Tغk?訷8mo8rF'@# n~|Gt{;vQP;R1)w1 3DYE~ۿr8Ɂ;,o[Ir2E\Q)':)UB/1RCaƄYdU|g~{MݓȥFCEH !dM-^K!`7{Ѕ @$5'$咝ǀڳf/RmgI.['6#Kdl<І2bIVMNK  $UN!Q1"®joB3,R I)ld() Sl@Ț9RxbG I`S#k>Y.x;9S Cdh-FZ>C_|Kxf =M`N f-6jr[-[QH,x:K}nCou/JmGZEeS& \QYE2`Eᓝ~s^.Ym0l{lGrt@`ٷ/RlF=*YM!/Ӳji•yLқTn=F" QI)H;kai (W$ mo($wՈQԱb:plѓ(%L . &}q%0h*gAʢ$ZeP*ɡ6%U'rXQ7fSΆ?-ʇQ/ZxR:|Y׺ZaE"ODtJEgs)lPR_e 0@K0$RQCO?13Eف8T/R VE4ZkR{#=Rt njyZ͠V:;OFv8II6)@0lBcI>$tȆȜ$r(A͟i2&35!z/rbF"ֺheJ)뤢k;k)]Dijj~уiԌEv>XVy+I"ZhK[aj$u"F#Yź)7w6Yivck3؈CvX:s ϫaMp{tb`[$j*!aL6yde͓MV/%Ox{sQFu7ֱÐKWpc 6 ڴݳjh> RVo>8=;zSd_62[ﶛu:>[ZvY<_n훴1I@Tt$K1[ R6eR!.1nDp6uI6AuM+ſ^GT7|<ϫ#[L=Jz1\6/r׍Wx7:;8*vxoMalO6_C5]fHm0=nlGDH p9+^nRǙWYv(5{]6tL<ҁh#Q:qik4_ bAYoDR:D(b`*fTvPh2CQ>AhKm!}m}|4 M309&mF V Zsw$q4Dhǽn<.=ӯ.@ڶNu/kwt:؟4 ݢx bFDxXm#Os `jֺ NCv Sdjc0bˁiiS/%$@&Z')"$HE c2_*c0| n);tR$ǔ)yeT'Gj.qrѪԎ`bR/팷oB\jEѺnxɔNa3Lg~Gv~V==C{8hnw~2W& 3^΋dmS݃lB?\ATl~mc x wp.2 dc{e'7=lI~E-)vB?vHV:&k9g@>^eN鷹dS~ԏ(wZeqٕktElok<&_)w٠}*mߩ5Y3-;Pv xc;K63T MqB Nַ3L|=N%nŬ'iMӳqeg[Mɷ$&eՆZ sSKم)HѳV lsdčj:RSGK5m3LMVrBJMř+q疼d8_}9M?|qie;f7m/eH/-㟒=i-7aZ9R[T{W+;oUr՚]%$_.X.QhHt5jbL:վEO,)$X[(ALhaq6!͜gq{z꽿l|6YS,䴮8(,u,rq4|x)$C,͋V3yˍrcSX7"3TsͶ?VK%SG-dOҪuww|{ ٳXIlLƜ}16ب9"DgCLw^ff枳{[,<-s"CjyŁ~>[_/M>;Q.y k0uyFo!p&jB{; V)pC9p?d_ÁP |X}i_a TɷX;+ͱP_*ZlJ;1o[9l!0_{{@OÛytN7Xyu\:afsٵ`m7w,#z!Q[jçŗ30ژly@dlMVL ^ý^v}>M*ѺZk}TEJq 6JG cuUQ9λC{0''MOMc8N&56L5-N2w1m6svӀgVCOKO? @gNJ8Y$}1.e ?* |hᜒ ^Ӵ!bDZ/ćҭS4O;,bo?UkF1qKvH̤v [,mA|'}cqqo_v .:9ws 1`J*9e3ى/jSd WK6tVFqɔ>^1(H~=JYD`%wQ91{0٤ .{<պY GO#AVnN3'=TQ+w~;;H%}]I}jˣ>JGGg<1#7թU{O4XbEV'p?9LCh\oh/w͝azU v9T:|᚟.MG11 a}ݍ~[47rJDZ\f-sqqz~h)?A$=sˣy֦*`y3_|ff?4ad2`pvkL2VC: @d a85PjcJfȪ2``hL %~ HAcs=;nO.^6[mNG,k&}! T9[>S'ثF7//)1^wχlG73(Cc=fn0OؿZ?p}ƦrJNّ +NnsU|ceOB!^i^[EƕxkUkuZK$_vU^ӈ~P;Co W*7 K5ٷDDP_Pj7Ag.ϻ.[eN'^+CMtPиE:..7 1X[ݛv}!EߏI$'v?hCnي[![=kW^;V.hٚ MW!c[>>s;{ߣ[&y:~^r[ و?߹k~Lc0 1μ>3q9d( ZáU-u1Y]uNBNtB]=x럫 5_YX>墖w[Zv52ym.%gZ+jhn*2]-iF.;scdMNt1{.IU(6>)bXӷͪЫ_ctlX͝VmbrsZ;Eoud9i_4K1K:Xi<Ӹd6IqjJU#be=|i 1،q X\D,Hbrx_{ ;ܺ"zDgd]@i!I|txM K8dѡ%MF+xmi,{ إ)GnbVmDxfm|CPFm]JS vz,VO @Tn<L /BAX|`SR'm!򬜑cr8-C`0H/ʀGح6uf H+QPAqvqPDhZb, Ұn*) wU2B F9!X[?X+,vNj$k&Pf:֍!mT 5}ꭨ *6i,e$Yn%,F \-y`uT`C\Hm0ݹ+Wvӫ3-C|ji/N溮LU@٣7669 4IYmtF m+~)wRiVsE&1}5Mq4RVK ]ȘLy}ӣg%토D5P%\͡jiTy*jK$]6p r&=E`CxtA)•.$[o[,?U?]!މ( (Bg6j0=+ߏ p1""4\bnV/Q޵>/#( y.3(b9sMk9=fn7A@Hz#)!8D,$O2H .G"%"$!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" r@Riu/=.~<׬k M2ǣ?@3u`椗T_>Tc7٧^̀y"fx4_UsF_ʚBځ.ʏ"UƌK=wGL9oo77l;K|g~VL={*8O58"UHuKEszT|M`-]Ix?(Gc!E=nBym%uEլb̬A1 :Ai)/ y!NRu%|QVQ#6݂}6RśrnqʌUe^|YN9r P= 12x(a *q%J0<,ô8fi>/&Jʢ5:JʢdYW視@^RtBGmEEp+;/!:Hoй ~gv-h bOQ({?7Pn=I#зg!k&> >@I WRr ?wAȷFV?݅!K3s/G?3w\ |HB(\|C 92BH*\==1.#T0*%>yrlBg}]B"l5f1׷j~˕\''\#{V322{?MVm4熲ϫMsQlTkr=ZF5B yP0Ycm]VW^*џVVͯy>Z[gjfv4Z+~O%D=!jKm}¼Mjf2&*zݺRYJ-Q\E˅tO|=2p Ju[w(xFH%V[4RPV)pp`7&2RpOlwi{p;hwo@[oա5:d{CÅ%vUhՒYC˕pOM%ނhfBGY+#_^.fiaK.s %U\n ZDQ6Nk]7bkbP,HB9q8X;LIuk_=vbߺĈQnQ(-ir7BwUj+Ynik4Pr4N d+0iJS,h̪Ȏ"C~rÌڿr((ѩ<ħ1J"UQV82dUcאD 7jOQ}lNK5g0բa[X+)a(7uVD"T/FP}Jo\OuR֐W[_N? NoɇM] o6 ` ޑ۟6Pe>I|ہWJjdPuIg+ _{îb˯\^] 3D/|X7(ߖ~)f\}0eaN^CpppΏߏ\ap}~Զڹ؏VI5Ό"`KM!PNt3iĝ:ܒ)*2.Z̡S0Ň^ d\MNҚ{0D ߗSJETov}{;|swvSHVU+sʼtNқ;i0V?7v| |0KJt/u XxdV=Ao+;[U(Hge~HJS0RVun"+"4u$?)SHޞx' Y3 S0^*'˝KBXuUv_5O'ƶ 'Lu Yj֎qL! /yIBd$&Go"i ֐/яKPicJ#KJ5L]^H^Lfb2hҤեg!j|YxrhVw7ɝ#0k$G>Ё=,' SB J!!tS%"Me2T_>1<:8h&5S2XolF8fQR&5P f}O,_Pun'W˧Tz?xyZjx@EܐN_W\Lf\Ms:pⳋt< 庺 ͊ڙ & 36RLӑ/sD@Zl]#O?W3,Gda ׾.~;)_f|@jKS6WC@Bt8eɬf5e[.|Ad]UG-._;r j%V "3ҵ)͙, CX͘Bl ,x& &%Z&RB~,Vq%UvL`eiŘR?QhR9<(ӎsd`d#eM(æx%dFHQ[%sB]jE N6i/8ƕEƑ(2ʰHH%d7it,i438!N`֖`%d PO:N[ `+t4$dFR}/c8! \28=P6Jc~QTtnccױ= 3oz'{=ul?cb~{(q E2bVGi)LQ+ I/J{aӣx2Q_/B7+7ċ '^<}؏WY~{?M׶T<ԙЌeV2(]Ym= 9yE+ .]8(޼@,7Z=c;hW.L9ҦsDP5"()*HEDU0͙NV%gUUl§驟BZmpD Jq^_8{Y1nPgJF9פ9Xk*Fiť˚9kdUf/U7jq\{/vwVfLӧղ*p/j/lvnu|eOlq}]BQBLlc4-w8Pn eB 6ǦQ*k9>xi8j(TM$#<^q-L+ًQ 83YpQ* Jђ/q `bQ}=Oޔ7[IR'Sޝpcs *V#6wՂS#~@כbRyvyAa6'T@ڈg!k&>_ei7pa*#}q<>g"f}r#W%nʤk L),(]lbx|ָi휱ۿt:UǶ !"n_H`)Ȋ|I 9GzL++lhgEMSnȞ"y⟳V*ʒG6-vߦǧ?I?J >ݮI219OK67{ٲ\mVa3[3E@6+jBl㮘6?'M$Fo)˴t((FHeD+:b>6Q 4]SIMwLAg ,$LE"ObYfZ߿̼ QUOW4)GX`@QBDdV ds7 VQmwͮoƣb&FEu3qsGji~6ŏf<7R;SQ9W.Ьfqe mRgn^hzI h3;10 "N,3}"P"o s) 7b FKw[aL?N!tQE#`NSX0nim͍-MRh f]([LsM? P?xGc; 0]>fUY窯"o7=&]3X&Q8,Qj<_~Ypeou 4M9A叻wXVl6 9P@7x s[r " qNyê*:Ӳ9Aⶇ+?lIIn}4Nݨwh1եJsT91$$M|Dy'Ohv!:D'^˃`8XrhiVvx)M3X 0h&9Ma1mpxЪp,q_TOMQt Nf[]j034նDu&%%SBp[~ +iHTHݽ# Bx,LϞ|2ϒB1ZR0?@RJ`, Pj6nc/D5g?.K&!ηzT(iJ( JI$9FM"cZ`BvlH!nn+z3HYը[/EC^~.[%`sh"0Pڬ͉c8}1 8Y-tSd`3 LJEPɥYBA"20K>(3a7g O b Kx)U"42)H+`Rк`oOp^pL W7 h\pؖ3 QTI}ڋ$' %z{n9jJ,f(U({du:ƿۄxz⏖W+IY`MQLHdNK3T nV m~+%FrEŌb!iCB@T&/ XY҉U`Ev|tSS4x1uDA#^DuG_`Q;|qfZdR0b4c' ׄT2ZP5J6Stqloj0$D )r"Hc m.ͼ3Hc}lJ7R6C)zG T;JF̽gf%B m,I]BgAs+f~Tu>n+RgR:jaCOCr<_$+4Puo^5( aޫ R1G٨= JK_[Nr}`8*H<Fy*iBr2y*X2cutrYm5Obj G戔>i;́-[Lr7g䓌ks٥]3Ǖ,fِ&uDz>խ|qb3Ǖg;q1{gsVx̟=o%3uѲK-Ac+Wp֋tWȫJ82?4]>* _9$#O>]>`>&ٻٳϽgkvJ.7euqdgLfRY) =s|pAʈi" 6͠WS+]p5~9 !+Bz(쪌eytHdw@Rg佒f`yU*sjw<v1=s\W:c&`M~E/BBExNWp0X޵ W"*x?2u3%"GӶ#tKZ-Cv>ao~v_޽"Z9ց=^Cyk+Kp<T hWxzx=TE^>}S<\7{D0tn.q.lVID~:(c60{{3‰:vru4QJ `<'S(CX-w+u|ol}n8$6;BPP͉Jh(t!H=rO8*QУ*oN|x{s3B-n~\/'_>iY9#GO<=ޯT!y}RtĐq$d/N hval"@;n?۷q(%{{tFXAuz Ez3ϓ|݆pkۓQ8iN%$fwI=i`,fԳG=t%%&8+ CL .dzܚx3蕧Y`F@yiy;((ښBf3ǧhL}?xB($2-rňHr*DjH;[D6-;vnڶ㚴GvM^MCBL1+4 zp T%\u⣜ Uf}1f/IA)&0Gˁoer0).ujaiәt5 "rY8@@9 腸B)IWro2 xV XP,kr$'d,Qh>z`Okf`87Y\X $5Gi |eMzѯ/Ǔ)j`̺'mVxu>iBd4)R3¦KġN|𜥷J$E!s~h2UaLѯv5Eg*e`uaʉ){&)e65q5<>N%DŽ_ h5bV#ŏJ!Axc]{m׋ݫqҪ1oX=㪞*zlw/?96*IN7Q\%LjVYkѪsjqVRHkh"z>F+q?Mj^ 7?{OHn_ lY ۼ`z6X`I`",Y/KeYeDVI-ՠnYbb-(\i9ht <݁UeS |j_ܘFC%4T=}A7 >#Hk _sDJ\&krwWXWIRA5{a Gv4@<@`; e{gRzk{=L'3m1Lz+ ]%3 &Dv&=vܸ :n8\E:ֿHVI3 $*S=*o\&9w_/p!בԫ6*G!Jf (R`uTanpŀ*S:kW%^`:}Oק)Dx@+y`y44!xKR7}X'>/:`at Bzq=A!kհ'Ȑߖ _Z@ K  ]NFj,BnDZ$FPHZnMՈSe[ɫ>|jSĦO=ᘰOmoDΧr|͘^#( oUIA&T7^ IUojJ۠!$=g5M _?8Ikn}ÂAɁ1Z6-y4m-yOP+AP՛fa/랡? !3HsUy"ʙvƿ'w< + okp :mg**eT1lF)P{4F7&xRI6# ;4Pq/ҿV&aDj QYJK\@Hl (J'&HCdNE=qiRJƎE]=㎀.<8e8 ~,KD?;y<7p`GN[˷0̌qC@MD\@<8d4mQ^mYRIx1}1{KޞO&%1z<ίsnq! o1W6Dž}rMbFz2pi8 sp}=lR@<) 6!Mۂр0ǩ(y 3†P㉻!a%9dם) 4k_?:$LdL]ҼoQg}a=f\Y߽e:3?ʨSRǺ`$D]!y*^4wB-OeÊsW"1'1L ھMȳqn}Qa*{Ns(1$ \uD&&^h|2N$ K|On=G Aa\͋Bhf#ZJevQv>km&}Ѕ UbrdcYt5kFhi 2Z^OnYFP&;BP f1&|Jש0P빑(W]>j׀U&7ӆOIѕ>ipLMt|FTXc\F̨҂q cGZR(Bc.ǻIU3gG`ٍ2X:{eK3Hk9苜˖ϷJքZ?̾}Q'p[ṣG"OٺExov1 !7瀪Հ':l n(ZpȋD]WM@qbt' ?c?/}ر(A>M5tdJ34Ukk ckJYiɵ~2WǙ#b! ۙ٤+VqӪU/dgHڒpl02$&ZYbVvK 8&j4U~leGjTE0q)7ƀ%H0qBˮ(.Udļ=N8OA<"W؅$1>0zJFIZDOoT<ðຂVPzQU !^;3|aⱈHZ]KGE8kcJ_*1Z#>Q-I|$8[q~[ 0 )<Q %ׄލPPrf)İd459=BW8ѝ&y Ŏ@;rB"`l_>X%) XD ;_(Gh'80pv<1pv!Bh1zLIJbFB.c AD'.p=2ut*K\MVA~O`6=($(,ol=Ѓi0M}S^i9֔_SNO7˗('9 LcP)_+Xp+8 !BrIP"5MRN~͸-O(z.rIe"Iǝ-(geQ%4LJP[~r e@.ǷQ P[/gH׌I Qp5ުSi7Qz20cφ9T Ǚ ۤ- 'pG )dVO2fR! ]Fb#0F,<{/UV x|a~Rӱ=KBQvlY:~T%Zp4ܸ(z' WJ|9I:(^ZUromRю^Ř'tMgL[0f6_qi|/ǕW={0l2̢IïpЋ5g B\6=ӿ} s5N:O_|`%V6jˉ?߷ [x &y$;r;ΉL!k J0 `2Ln5Rpbʼ !J CȰ94fS1T6[qbQFS뵌A yȍ>~.Ul0д|,V{_aT\ F6zl1guVc1lRqqB" {H8g=8מj1&@Z+zB"*?n đq IՁkV#@w(v7Ry ~U31N8m)pxU/=0p8?DG dJ%96WrG-2^0Nc!*U=$~_}7L uh1Ruh刺GG?{He12/-Uf 1VWKX'b]狏 Ls%@ Fk+L5t.tN#(VcfT"Ƙ9I Ԅ@ɽ#ʋxjZeZ^)v#+KnXΝc`d9Y+IQFɍ9Z)Jx!nח@P:ArfFkMs+<:!]]D "sf <)?K;F]"uQ"*2͕Twʐ,(S`h|qT27PX<3srۢNJ ‚jj"@,[ sS:+G>PM[bLӘ`"5*i\bDxC")s[:٥pXrHW.u LpE{ոἨa\S[eq[yN[V'Il^_ε{-^|򲜒7GC|䔒{t97rUePZ]s'k \?fT|ba#\`ĕK ~LbB =DT;uG+$ydѕ nc]N "8uir9mQ%+j%GGXcXdU.$h͈t圑$sn ;0V>V^X[7[Yse\Y|tFYv!= I<(ZI" ;I9@2Y5Z@A5y5[hi,gX3Li=UH7DRY59N\b`#k)dƊq2kVR(e WwSk+Ӛu`ckn9Y7H-ɕH<c"%8Mv`ԪʓA*mab6΂ ׊?m%P黺 r֦'LB29%1A? Lxކ(Ҩf ך /ut!j{Afk6|^1k"19"ɽ|6_b<+ZG\QN":ff>JĽ<=?/NbvS㠬xrb4KTn؀:m9LSx˞spfQ0۞R129KgKRr^blJJi񨲚'ɲBrN@u ɖ̛ۇx[RTIEu&RMEIeΕ,CrdFC}*ner;vNYS.fTmlf{r0|4`E/oXŔ&~Ga ha<J6{ qmuvh9 u xz9jZXeįJįM\ lh ӨQ᧾I1%T;Xurv?|3 /#[;0~Ќ8=$+U̻9Vy1*y1jyl%gi*_^Qdo]K=' Q]u?\ `o_~(}N85Lbsp 5liXmixm)|i!U=i;i:bx ɾwE1pkqFC̓bGVeɥMHj_O"vFW̡ lZ=e,вI+k-ɂeoF/Z©T@A-DzrvWK<;݋ޅaA$Nnb"U&HPE (oؼr)NO;wj;$"oO6DHg#tmS1K>oXykfif.]yՅW~f};Įן]xͪ1@K)@RȜTd,huݥG6Sr@#t~ N4tts8U[֜QYblD*ΐ@}r}q<1[6+bO`= m\r .r8@pnDw` qozv9aYBc#Z,Gb4*JՉ%j }2J _ -ݥm'Չ>]Lx9&d}l^N$_b62zfa3kU9\3iM=Y݁AEw-Uz+C`Ĺvӫt;.,S+3'ɡHqx4่p:C̖_ւ|Z#lsZٍ*ݪM.\z%=&.D rJ*@%O9فj*kp{3 6gNN j߯V|'C䩍O=[_M}>>’x?ut}{ċi c/fH!"WůmbTf뵝)ӦVFOȬ1:^R4Peb=\L= _=d#N4zʟzߴK7t8naoT4;0>Fcz͡? ]{aL^ߨg8os\8]vw?->4K*Ѭy5v*l.`yUŜ?O-S)1:M>ΟgL/y'qIyyIyd5FNnɂ4E堳E&@ :gN!(lGqoA ʊ+b5a q]Eh&O %/(iAd(e Wwg#˧ϟVIiLOJczRӓfcO^Ze7{H$O&Y!cf/TWqLy:fR)x+)RNkZK\"fO922eqP0BgUzEn !Pɒq踖@?Z:r2*1Tw[4? ҈_ K!EQwəLyanU:'wYi]"qEXf Z9*]ԞwI> N@V8݃؍Cl&ePS {P7֮ *u:ⳠQ!Oՠ;,C$wpn6Dl?NnC7$ Y8i A-[QIR,i,'x堘-d*O[P'/VlLn8?4 7{G2NjmT'$ю02r a-8$74nӧ{1S <Ȗ2j2o0}셾RMߧ,JEqboFKY$<ˣA&Up+1!)>3xt{̡V΢˭}?я% o!xq_b񺸥IK9~Lh6N/R-]ߖHd^/VN3;+%ZJHJk]mo6+?mѫ Abv{D%8c)[/j([FŠIs$>$9BqVG1:4YJYbOoӐ'YȅvC:6'"#ّZ] yƞ$m0[(2S(4b($ 4,@:Q< |6" '(l+҅]͸l_$fLbgD4h$jx w#g?F;P(mzYe,w45pX/h*-\h1Q C)W4\qo3f U@\h:4"{zh0 kpLfpuYجfO fi Yfv{Q!M ei:Dv+xAYL[%Ld4TEokX6K~M؞Bב.[cAY!}&'xdߎbF'4Bpg!D1JIJ2"!LX'PS~mi/.\E9`c9 & fnxcp._<5J{`itbGD04 # kr? 74W44T `Eu OyTɫ)l(0A_oĈ", A`2[u{khC^d!|S2V 8c؇֠4%[HQ}@=.f|*۬=1"6%dWtU!)^qtr졆 (z#Y_jej+35~} \`9.#o^ A)qOBLͨ&fqHi$Ȧ1cxW+4,]ֲ> G3rƌ`G*njwj3[!rrM؎޸ۡ^X'ѷ]#ϊSwRO0&ۻnO۳ǻI1,&^mZ,b{+"'!^:_l1" Cb/P:o /8 'R)xgiw{P-G<]Bt=SyB;ui p%.U/Mxμ. 5n!ɼdUSjgӺ:8JCHaw]J_sBO܀iSx_b"6˷il~8 "1׶#z7O>D<$7EUt'Z!/8~›p^)yzU/?O] r7kR-\`Ofqb4cohI>FJf82D:&cdX !~67i@a=,qODSDДi.31` &i W4Aii~I5] !SbAmaBd2Jc NsZi[`-O[jP_Ӓ6"2Q狙6ڿG |d]c~=:fRwß&2ў]GVWۚCy7bOquh/??ռml2pRMzKDnR#lO YjE>wA^ַz]_60챘?ףLz[K.l ]+$*H$(y+`S_ؠd9 m8B=K] ɍh-sjJXzi#KPW`]J쉹zO&z5ǯ8ARriHlHDH.ӡPX0"X{[՘|J+HjHmk.qf" k4qrE$I3di&(1 O0+i5 ЙQDQ"0T((9Z(I)\M/)6~EK0u5M4MFGL͵Li32tRߒF BFb,"aq&*38SQ̌`KIdNZR)j&,%ʡCibl}F Gwܲ#󄟼B}^h==BCPfZ(XyiPQOmO_ rQ] ufiڵ}h&~&]s="ѝdcrCN `TS iQ~;S74=6kQ@)ב9p-E0'MC<31Gi,LˤM8]FI2LK[B[DHQ%8[:'F O)H5k,PKax83LX@-:&RjjMck8iEmu&Zrzw)ףfj?$贸r^Z!݌z&7ʶíJ&X;?Km5D D muHox.>"+Ơ R c3-0 a~V%wmhua_wdxbҋ`a[<ֱg<!. =޼ Y@k6N~ïaw PjSq@M<?,Mfɰf/E\ZRvHXu eQ"ii/5%uG2C'vpl%w owRxEjҎS:c({:b*qaxM+ (;tۀo'?T7GC>b(>{ܞ 3`Wn`WeBH_qvr\"qKb"a!%2A%锃uA~'FIƒHDŽR[dZSB4R21WXE7v6qO%+V~}Y-ռs|m\ϳ3ac*+^J`~f0|PeXäERա[bDaFA3 (hraZI-Z=8+ֹ:#s};ۖujQG Hy*X]st#Ck 6OR$]9Mz0(TZ?)6l$Q>0Z]kݧ0tm$'TS\,M3n{M wksH& Zban(:CEƜj{PqNHXC%֛ͯL4 jfC4/$$篅7A/@U?\$ټ,fp7iɖdYNpizF ˚i6ɪSbըۉӌL5֝@awݡ&~ 0s`F}hr$]Kڸ_t| Q$8wyxûՔ%-^}޿Czb40B9iS 'Ē6:ru[H! N|*&v^isH9O㳗~|eT୶ww[Bbr6pR)"jH5 V*+BGW1ꐥt:8!Դ߹jpC w^Y{ ўR[ 0rPirhi5@3u}9 ? 3'C$ z@ވturT9dM1`EZLtvlYU!:imiz_{*a^;HxȐ {5a*ؼ A 6Hc0.?n^֬1 @4ԀçI:[m $L} ]/1.¥r Xð;0le09UɦcV<3۰qdTAVn>it$3l fY̩M{cqax$]vA6,T0. +HN ,Yx0ADiGaLUc%_,L}0QJR߼ʅ2p#G'bc=吱,t2xna%A&X%' rֆ)I; c?B>[ k|װJ*]YkާX)p7;,dP!;0Afwr HUd=*Y5#U9˄n^z| +y__=YzQ= VAC`h~l'"@  QZ :L$kj6*\MbtAR٫V\@W#TƸ tk"#PJ w tM*B9h쁳(1r16Nۃ&,r~"wT6p8SP~C jI%m<3aW+w%?>%|!Vr<2;gglntmŞ5| qg\O=0욗Q8)ǜ [Ƴ8#@Z,CJYP$iB]y>d^{Gҭݮ{4ר%mOw%@.J݋ssV$$hΦIĒ6. {OqKB$k'4$Z;/džյ:dz1bspRf"\ƣNƆB2sjvuap͇ rCvt.f,/{R f,\Ďu<|[V^qqV]Kx‰ع` \7IeܱrY9UHkyK-yy~?H.zCC`W5jd )A7w9:/^#w+=31ezܨ\!k2WY9d^ck AQN>O銗rx|XZcId7uBuc:M"rYzF%5rcHPfc袄N/K(J;JÍ!5Iʤ ɞeHsH)^4$heiGSsk764ichj{H%X2i'&vs쩋pvXv٤)]$ |?bJ2c#Sζy4ՏS4;6~on|I}Jy2LjcV7a{xhӐ`9iw;]z|A?{bXbΖ]CN-dQtx;|׿rknM/I 9?>@Ͻȇxp+˿r =h?߇4yJlw ~98=<՘o X/y-4^Dw i80ó*NpSI7(7f7vJI>#I6رxs-S4;VJrpd73ZЦW*+$oy&Z3JCcOcYed^NW<2k˔>A:y~ cx^=z,{xM[K2Zm ?rggcIz+믈I <ゅLz 3B3ąTqVc _(ۄ+;Aɯmla+6?nWsD2ٔOd/(M%g" P!S#F=fUd)&79N DA}&j^RL騚cOӤIKC2&ȱ@r-c% s$4`rAKc@0\r0`V(&Z?lZOsD& rP7]|m5'p 0 lq7 a9г6u|aUٔi#6̳/ͩTt4_%Z9hnEC(`T\JW%,CX,S 9'ਗJRK.hz, NpI-0cƆYxjQH5[uv ťhлMn`%v3ݔ!Izhթ1IяI~LRcbxLrܘxp A !ngД F+?)kJ=öl&hj4\jkVg޻LmCN]enzg\˫S`K [0Yьj;ػQC5|: ,洕eeۛ^r CDpZiZ%5@Uw'cٻ%+ּ4~̑èm0w{z4>+[ U{/+VDFFbl!GK-sN\6>]+&~>e1&m .Dz ;kZ#sr-a1֡RAAGCk暡iSo4,ԛ?|6ǸEC!猹VS?eYk5&qU%+Z4:6!SIikq#a,  Px.yL clfnB؄ ,87f [EXԑm&[gἷN8{Tdmt޽<MiSo&kuTQuV5+2lQQ2l)H$GbYH5Wln eGҡOfr"rњ>yói63IlLLJ)&h03<\E4oz\CV˽ ߾cyITk%ƀvo= ~w)u'Jz@E{ ie5LE2Y܁kIy(ݫ`ĝ )v_ t#լ'r֮h"ߵ+8bf*8&PHu-kBL xMh>!%NnYIk²h_5&8z.;0͵^̈eXX|77m0:Ϳrzon~̦dDƆ '<1>C)'h`ɁԷ61)ݭ4kBҬiKԕA Ͳ+\G҆a:A{4`_0aj3[ql`=Fͺ.6]&53g>~ 6̮3q1evo3C*4zl3x<̾o7BG:'0'tGDtu3'A~t?S'vLn\ƅ-{-QIycfv'3řQU$xx+ŠG˙vy". 9J*C:5 fox"3{7M@*`w&Ɔc rɲAz_o_dO9,W缘LBÌ![R( 6sLkIn.p0#6!}@Ե$7i3<<e@xH›i.޼:ń7CzޱZu%-tDb^] O30D1 8HkV MhF h$Sl4$ fN1Vh&A@*bJҦD.HN]KI2`/]?oJA]N*"MŚ>;GU;5 !f-D0]+ښB1kAɊKqnduܫ6ߨ:p둅l{~Ԯɶ{רR>xC )G0v.㍹p$"!\2 JMAc[Հ~锩 k'vPV.Lb@-We %]!}%' ؂vn 1ySu:v et׺Ě쭨'5zbsr%ݺ+ Q8! '(I `µ%4hiiB}}[ЂlNPxLNQcM :hЩFN5t|jKI"(^֑! (.3`.RΠQĝ6GVv#3LuJPf$ ̭,KI&Ƕg}9=z=زEДwӭ\tLw׾F ۑ=|>ɺ|aDt7QQAyY,Y8WN[Nm%?1FVN5Te97 JV^ 9z) ׄ*ʡB?YazRxgw픸:,%82eW<rߥkInLO)`?)1'>YB+C]\pLPh7' MF1R`Noo94STޞ5"h=Ş9n+68.Wܼf4$pk._<}:yMyGڴj>=O} b1 EUus΁BNmp!Srx; 9/Ɛ⍌۪Y8cw_Cz-m cbHy%~deN}=N'Y;_MQ)бogߡbnPhR"0Q;B(TW((.& LhH+rnڨ ZӺiClHaV(o>?c^ ~?.Û7Zy?{;fU}w+\ M+?XEs8sz|EH<[Iȓ&c ͞l~ZڡB/P6k{`A֤zu^!gX7Hfaݫ-nsMjДRo&ٿ7cgH&H^6D5~˳޿'x!uJWL/#5_e)mVWlZMN Hkꗫjb*Z]ES16]r]SG0qE$k&-3b!G( .0{bfn;sgn;ݝ_+Uo9U~q\,Q<ErL=hPJs8Ka,JOK,}dp>u ՑPꮸKP՜H̡Bjz=E*2ҴuGь %Vϒ]%Ϥw/1g\5:RQN JT.8eO(VFa̙Ԩ)iF SMhӨQjn ]u)q7ZOHPrU_q0uto?9tb\FYw'6IZ 2jScݫ.?E>B)]{א&s2 "'ӭY{ "7̗ $%~V *7 :Zr#/YR>d2VkP2^kޱpXg('..1ÙOσf&;pzPlfa L9Xw0*vb T(KdwL? U =2Ɏ@ f/o{꽉I|!_$&>N٧IiDS OU¥*z)ЛglEcǍ{ʬ[=Sǜ]޳T>'\E`O?Q_>ӿ7GHU)32MiʌLSf=~>r3۸SĨ|Φ7X0:"7!b2\:d}EY)^4 HU;V)JԊTɭqhv=D9vg-79wXy4GpFLOW[7Sm 3KxXCнLyf.bm>z0ũ֤|[>jh~MP$Zu!;,>SBޮe@KV ^ !Z:˟VDqX~H@&>?J\wgiϞ=w@sE5tOcl1RsD@D8v~r3+.5nO%m?_ysEw3FDTU^uEPBuWSO8{[+^_J9?-Le"RfMj  &Q ƅX +ENJ%ka.#_RN=a%[컏ja> E@y.ꌈg#f jn*,)x)@t݋as֠OsqbDC9 J{atħ<ɪK?1ӝt6~ G_/?uNN{3sc%twߴy$l蚮5y-2d&>6٦d@Aj^l;U6=Zwؿ?X>f6H1*d ޤV*i>-M9WuxݼΗ.*c^PN=(3t! <(~;h+-o MW.2jO}-uoq:dW+C!c.P#Nߥt6GtJb=Oľ#E8Xn$9uN SK;fOL] v]V?]C5*G2ՉC8oŸݧJcK`)+'#m r D 斞txBq7炓$n)ҝn`Zd,/!9!2n<HѳA#qi#Y3.Ry~^F'u^K:c>nT-Ypk9}>I#r1)s募f|cfpJذo?MHܽL}i[.Rlcq* 2:UHTхN7 ,xtˈ1lyݦZ@-堵T T,:f.nYmX`XPz}'d> i11Kax{v\8*O!|[ @o{v?&ey,!M<,6}qdv N.GZݗfnDٶc Mrh#(CxJNH~䁳`ݖ$PXt9$.vVΩTv͜0d7iߢK2O 9qbq捧lyv m3[6??sW?;{˗M;kCUŇwg "G }vU.>ʐE]vFc./NpLfU'_#*}-{?mSs-$)?f<ʔL?\2}H~jV"2o}}~-#疓y3߮ܙ|nq-?^}|yݺ>z>z3摩}5`_sa]?xeL_꜊\|NJs?~]L='{ *_SJP=$6&9hu]joW =GN\RADè݃ZM9;jaqLavp;!V9B{ڷpnUIt_k`@l.Mc\-hXjWѫ.6&{4[)5 Q"Ҕ8L-P=T,FHm/)ߠUjhO-Z\QȉjtƥSݡ!苷<ْ;V3f$t'ز;KS^pWe&X@'ygKtn8e]CCOv?j[޽Nn]8PVs)C A= 6jH!WN1L))nCr.XKe5MF_{-ܫxpj31L!(RUPzaE. ׯ!B)Ii!+rB&wW s^?`n?QuTKoگ ƮEUOα9nS%d[g9)q_z͍g}g>&%,I}hN佖oۣ㵽V!΍,y4cY`WcM@&鱰t{߰1J!q,(AI6jW)bJjhSiC7AY}I=1UI9T13)YEB"Y+'it4-ďlѿmHצpNtRXO )DK`vJؑ"S9!;bΘp wվh6>@{.}_9޽D_p뽷"xDf̈&⿦Z믛DzdSQޓaOIi{QtIY )rd+smQ2r7rHP,G+^:icomF;hVmzMѐs_do$9fM!iSP!*tCs.غ!HRثLK0 &l1X8_U{XRFThw5 [sM%zGqa2XkOyf͈Mn) cސ1!Nj- ոv6}dvg;t`T P)zC-i|&Δ1> S-=L1!CL <6ȡ)L*б%cM,[qtl_ϙʒ {Fue@O_+']%_ =yqq٠^_8"M]:w-$-IBEpe]RE]o#sWbm _C;E[4h~!w~K:y%V-I.H,zE >3;92&0sm[Cöղ#.*Ltގ:-5{ZQ9ucHL[6 .EJ乱m6PS[mh`θf2~yuQK<j\bneIWₘ+Ц.${ɢ 2(Ov]hgĮR,Dzl9bx&T%vQ89ɾ$sOMAGđ&/%>.)ٹºW8 v&h_Sw$j00% +9"  :0?C}PY3` NmFQ,j[#_'۷{1<R#AN! ssBG&2}Tp dMcUpTh{5Gm04}F+u! UM]D\c2\Ǽ-xV{Ϻ܆)2e#|)cv#j^hx%L˩f[ȟE=g?.G>'P9e  0 ɷ1AAY^}i$jوX eFE".P"r`MR)| |VrףؒxR2md |qhkJƘ(A  dUf7ʛm`oӚw/wprhl[ohtu)k̹Ʀ֕w:V5S[RWHI7ٴrq#Ghkh5EMhYԃƤ1 `wx$K9,Nڃl '!"O秿Rk#묔]iU"Y*1y[àEm0&uB&")VZHdD%ߔ FɓT$ysȞAM}9@ ce- .$dV$nA ձ/g[fL5#t30:DjuҺ+cn?[}gCx `"|inK:Q/o?$ދF[т,4٠"Xn@ԑ P"#U#tB`}(5&wjW (g6ϲfdLjS/q=ͣٝAOeF}/sE,A\cm+_8x+-ݩg>X-ӣ{f S8nƷnDnl(r[8q{}G^2*$ AGLOF""W>M=I;عmsgj !ߣkqO z:{Xv pSWV9a Tڳqi;oߙvV{[X=&vvnFMVj l2뺶FO# :{GIŢ\7[N<l)Gޒ7T\)׏o<vbEh>S;X\dFndǼ,z/ŋg,M ry&kf옳T ƾs9BO.ה5)Oɾy:PkM|I88gq4C Vo|^=Ĝ̈0f{' Co׿ռ@D=dEW{#oq$/x2w@gӂ)=`A;easK Y&y3$C4z PfZϒ3Ž {6WaG  hcJpui pAM[$rwjjΡk\9yœWWJPF_PݡS;QL-Q;`04Xcv@#:vڽjK))W"wWX)fCv) ;pj'?4^#XYЇy^o)ԮP當oTmT:9mr2]\vݍ- 7Y5ϳOQl˕QZkʯg^D&#>5i䁟¤}Us˩/}V[H1~fe(^7SߺZK<$qF+9hM Mj14Z#V{c5ZC \l_{$0`{H4s*UPIveqސ{d=V{"#;j䷕eۇiٮb^~l|Ah8Zpv{v! )S"K0gE*,P;9cPB oF,N7 rysĹ} LLv59u"+4VA%Z%{u+C Q0&#{bȵI`hoi)rb5 ;;5B>P5v26k3TW}E,t$nκ"GFYOm8p!Wfqyn܂n<@^63ʻ}U}eecA,* Wuմ::gCZ \^3B (EƑ!qX}P\Ȕ{[+F1Av~VÏl[54x]ac[ig6hL뢍D oyK@vʮb^. qlYCgj| %j|ŦɄZKNX7E- m msnD&Bڡ"'R{v_d'ښm+ zY$,*?rnd n4=jJBMOTҲD8~!tB>gF2?|nz& ] gm>k*{5iNH';5޿ud]^m1J);g.S}7(ˣ([BVё2x&4ټd_]=Njĸn磌]՟si;y7t47=ĜX7Ʈ/?~m6 Kq_d[]u߄\ܗk~`6?YJћsV 1Dk.o͉ص7RJ1F((@ o8 ֐g%{>E-ө%ϩ qD{._^3h0ס..УFcn}+S1$< wvo'N#NVT#JhkE%NS4axi{-G2wadV#mKEО40b (J8w'.eЉИ{9jF6a?Lpsuu+q5kҾ<`3"haYS9u||\5v#u"XG&opy5T>ӻnFM>zZf5 YućFzp|M u PM(4ɇڴ!]2,H:]ӥGj{M8{~Eؖ.(^Q+V%Z2*ZuX uS|m6(hz2hx!T辛4xQdNP^;{QfSWMhU`I3;C9ޣrFa _/Ds?xL@t.Y39e5%0a;H-uW9@e|iU{>Є{p u~L Z!rFᙴM*)L>&׏˓EyK6 LHm?g6d!zen W95 *a*9ߪ\ObisrZiEsq߳CVGF? 兜/i {Q2z!{r_rṫum࡫]g/Oqv5 m<Æ3b/% ;1+Zy|:"\gd{,9b(#eQY]r c$R*'/,̱rΓ5D%ǰc9]즺ni̍,g }v"Oi1ӑ(u Ɩ9Fsqm25l:SBx0MMs<`lזՉh4)iġ$=!ݲJ8czGqB4W)="O.Eٷgss[zUn?UWvӡZ.u<Jf\B# H*gP؛,` (#ѕp<_YJ:*qq9= nbCCq L￾ ͫQUW;}Q +K.Y>ͭ{J;eƃq;EMqF9vM?^(ɏ[SҚ\]E7"j~>9<8ZJ@uFyal8gߺX>aY+92E22Vۜ50W 3l+=)D[uqUnwUyRTV+(o{N;3Ъ>&A[w+ ]W x;,hw*la=-K г)1%-To0됄)7fPǐԈQʾ<& zyӰ"cbN?(^soH̺5JcgqH.zjFP@H' UKj*q?ޯWm+UW552V4r~pu?7?8Tc0vD$ e_ 1<]>0"SH1 ֝)FA1OFw1R;z04hc^ފ' rAg{=SlG1zgWQi qW*6:;P!MGqk˴>4˙e@cAq.AbMy5yN@Ġkdu`hĨ-\e1kFAC-9?<v'5;ͩ!cHSLhུ褥vQ:V-tФ(a\ y[A3#닳Vꛗ_t~Eis|/rt /"eL@?*ĉ›Bpc.`h 0ETև eTB~g a %-عD^\!o?M.ɦwz%.arraKWmkA}]}p+3˾j"TCxKz9ig$ÅE3Ԉ-dR, D}F4Z/(WZ[o) -Y[Ɖ C9PͅT9 *©CZ'Tք~,L}Su( ow%'==m^0]XS-ZV^*Io{46w2=>v Gj1s6q_]*@sʰLtggx6G:?~A&.O%6{,>$7j^V{ZE3M<2]sEHZK}H̛݈z5g ܃&-ccI7K pHCYۀIuhvO9य#L5oR!"8W'HLlpI)G4MQ)g{,rF;JgmQZ1)6 ?ԧliA-}^}Jq]9P`#V1~F)C2nG~nYWf-:戱 ήNHE(|ޛl &/+|2<%2BmlG.8r栥 %Wàl9W?=%|Ftt ɭJpnQ㝃dU8wB1mnr9hMyqXDjE%Cu8]Xǘt8%ZB @?OxD]I@f7!m^eVt/-{9\'OXFQWڦjAnN|,;ѥ/X. jfW傺7"22 &RjAOĨ逞BRTX(rfmg;OvQw$ wNjA KXf9),g=$eF"7y^$-jAܸGTuu+c0Ղ=Vj?AcKw.nC,A <|m'XZTdu %3UWl{{_FL%Wp"xO}wQZ.!c 6@y6\C@}t6P [xC~Ng_}צDjlal%Gt vJ5e5悎)SZB=W=!8'zɓ]ayl? &~qζa29N]tgg_pw㛑P@1 ~ nEvb Ki|iQr:;] <.rù=^ 7M.1_I*lʪpGE4Iy!G-dB''Ҡ'N~8IO''G$䉋2e e#7 pu,Rȋ1DCro=TiPkG]>CJ&+Iq?,J^kM8?TbtQaD0o-+-uB> ?"Pa]A vތXD/~y??~߾DcLdujw$LB o9em<9sW(וms3ZRZk.PGY9"x sG8A EP6phhQ9>sN7,|CqoNPm:>';½Z9,d#'$CUYէ`wU&`zXnmlM}ےd]N1fB1%[9M(ehhua> [j6 ylC{Оs՘'./x Wi))nRE68W-EIj/Y%\b7/ʸﮕ%pƶ)pZ/"Z@w-"eVПw\vĠ%c5XCoP ?p&E]Fʗ@I _h?:l8gz}RCX)92FȀRYFC R04ՀqW0EUt6>aG(<ösv! eTIJf9ilx7gV@ tFÕIdaOL_ Ə}^`ז1k]Cu ~F3=}F 9?1dҎ3w=sI. nvXui\.BV}ܾ]pZL#g9Z<7`mϟzr/w5WZ,0ҝ;H+Cg@򛱍ѫy/WKM=e{W]Ljs0sš^r&9H\[q:|QMߝ~^fc?/]ZԺ bQGY:~b!֤][3$>Zat+,<k\v:|w!g1tɘ 8kFD5{NdEaO>lӒMrZt7P ¾QTsQq}4kVEN=$:<"j8뛘\Nuryflxy,+?NJW)JJmN?6|,WbK>4(:7e ^uvA␋FeAtlȕ'DG{vEwM4pMJ!Dw'z֒i27Eq'{Qp 2G}5xS]i2d$m=0dZ+jCG.I>>*p)d{eН> !4 hMm \B\s2:e XQP!(SL , Nib8虚F:$mfDuVbw;O}%-{sA{fB=;q>L TX+[}’9!` ~X Ta6 g2-hfQL`=qM}teҠ) Wk1FsIc VVF :nLqgDuh*X=/Ke. '/7yk!#tQݻޝKF Z )Xs dgq:1T?;,H&'3J3,"3Ju*+.2z-1n R;V\ɱbEOgrzQ%ɝ\s~1Mbۮ߄"E0; ީU5p'D I;J`r܀[XBݤ9mn4LÇ5Jq9͜}P0*cƅL1!2gy* |NP%S«7ӓH;KGy}y'Lw'Z&p,y:AxF3g{ RCL'hIJ#Ngƪ#BZ4C" {V wbf49ZZւo W#vܣ~DdWkPGݷU NxZޑỌJJ@GK*YX2Yn6NDxo:ަ whtjQAh<6ߴw\4&L02Rl}|hT2=uw$"]b`eԌ2+{lXm]j(@PO;4jw\Rc+B,9VF8Ju(xU@֕O/JJ(ATIjl)X*w9 i;F׈kUwgS`F3`FcwѠHƈTcҙ=ЗJN΍^/bMdo{?/&Gh=1oy ,Z<5&~ұY6ƭSKOmSϞ{vћ&|P7MziڒqӱIhL.>ʳX=(3joc+PD;%={Y.k@>4`5$DX0+bʫn#(1%fx ٟ\JYCJ"m<b1tWRNܝɏO/d\Ѯ}8ke88' 'w3Z](y@'q֦@g1].58ÔS Sg *T׎=E0*7(&+a|7f _]&s*7>%4(X{I/ut4OMFiNm7jngCo0YA׵rC]PR+|Z]=JM20.ӔqiOyCT'5\BT o(O{uHfRsǚ`N0}0gSzuk`= G&_rI^ )4;:ZPxh.X?z/m kpslQ9hl^3u2 Q_j C_N$:N$::.^`I\iO|6PBms)SHNIS &1T<k&}s]5iCYjc2ƅd_͋6 i]Ʒfo}ֈ]p=j^fƪ2̅ :^ :^ ՙIYam*7bohDj=~jǃ?x|PMkkb,GN%T2Uc|b;iC"5 r:asA4:gХmlheD+۠je: ,9&p$lPCΦsǏk!ѷ=* m$?)qTWlzD(P=F/R|`銎OyOf,Vj8M1g+o^Tm!t߲dW~0Kb#Xh%h,+fxdk,`S4I$PuZxt]bv'W?<:o~0/{;1t2+(':YŻҒ_/htҌxiln2ߪ~L?`/;mץnޞ'dZ ZGzzX*r ymr~bFd^2Go8}5ݞn:ئ?ξ4`uSm$,A TK5% .X){h*FTtGdpRXȥ7E5D'O0뜒 MF> I˳'6֕{;6[z~ǯ) :}=-ssC3eo¿c _-+ſEGՌ}@-[`tҕr1Ly&A9r_&qc-P޺nkF^ GX,sLfO90ƻ6ڝK{5"V'tĕD +FXW CȢrʗK6c!4 KKqrvrj3BP X3~79['P nӱg7;)|-dgo^kPסj&ǣq +x\fg?ߊDO;{Cj*Yn?^3_tX|,eof4 ";M;3>8?N.?>~f< FY1˹X\/|Q> |L?*[N'3Σa?zϚ#ja ?:"E9$.|tJp3?75v<~.ʼn}YP_g/y:4A2ۛav6e$\'NpmPX9N\@=^뿾_z1{p6 Y&vUׇ쏗F-0iϓ<^F1^wpWX Yp/@}~5o~{ ?}^F׋ ×P$7T~~.(flG 5c>/ۣ`?}g(4;]U*^x*/Z`:4g)G<-% Wl-[`wWY7^-kvp\?pqyߝo~`*<@?OFC{qup^9e3OT,yO@zS=~*]0UwXD+\^Ea6|CKO.+@]KSb\'l15ѱTG|PuSDrD28Θ,Gѱ QuEx q][sF+,=잓 _T:uTl.0kIIv?=/ HJ )UN$ 0_wOtOf-kuZ'x`jϼ{w~rp(gzّ~ن $qҠaP_<`=5_{0 +{(Ռo߾egDsiL-%*i 3ˇx] H7d ˩҄ D+9ij z^3h0¨sv[*oǃdfݡЂo /p_i4ꖯZ:Kq,:FI45x2p$aI N39_[$˔|3+`u9ɱÚ,m@*Ϭf2]vLSoROm@Qf.N#XnϨ;&BKF%xj8rHHq% 5eK# #?$g!ۢ˕.z-BoxnjGk*eOK.9Vht{ccT;&AbM( !TR-{tHkDXܮ a*)i?dyHL[3Z3cՄTU 3aO"fZE.zΟ?%>ȳ!8>>; >pgߞތ׷}>{21q"V .i!*6J{ƱE'FU-h5qV\E:J\ϭ a0ߋ?3+g_}}= }>(xUBAT\O9¥-u/W]NLz]Q~; JieRLWG|uhj6WӇ@+IQgG(kcYb<i#<18CʵaGg|f+~b#}47GV[B{BFb!\(QJO}v}fK]_Tѣɮ?Gd_LNvɮ?5>Iz]yha%h"SzDK"L#J?",(H9%[r,m~TQRT3hGijJ~ٴFt _7E*M8y'o`o8y'o70lO}ji(R#Z&hb$J)7P1El۹9w_lϲ̲ӵ៪Xqt Wĺ5,ݚ?kMݕ-+/e{8"lH:noB2U0-ewX)*vX\5w|Q]:rsK '"<-Sbilr?-Ӗi˸Vbɾ\}ܲe"k $Qڄ14!(@"DTvZ#sY1.֡Jg U|(Joy1\7~1Ǔ}Q?ӳ3"(-0}h>Ay.ȃ{"oh[Mm;L =h*U't0RλRր҈hV'Z$:x"e:!Rbbhd3^vym<%x8n1uddX8cL9 MjzD%T}r,ub^,DSɨPOԸ*kGL))`RZ_S].jXѲwE/Ge=k 6CJ,3'=2H*(6KԦi4qs5ˌsLn ,qٓYa\-.j`1лaRmLJ.xCLFENFq{츀6!D0btu6j^)AǾv-Hݨ7>e0gKggy1Av>{yIT<|䏵3&-:j -όJqOU5zek֎j0)e9p(.pA^X9N{6}(3̂5;ұ&CVzC-pF7i+D] }׉|ק`W Y=x_s&}Iء_%7O4HWv?lɠC00KL8{0GIZR')6a\DIB9$S K.1/ڸ@.%)tQnyGRz烢<.VQUbO?E(Pg]=QH=eYaÜIVi[$r:XI>MkSK^G.çҟ Ky0'`N>rrZI uO:2Oi[pI0a"V#'sgd%+qZ078 gwokSZ4ʻ7ZLmK. J[\Ӭ.+~EE>p|5zʢ?۷own{^E5(*zX4gwbd~3Yb|SKCXL2Hjb!R4B0vb kBs94),Ra<+EhG]cNiiy"=1*=9$3J32T O53Oas!Kw5e\|zGke.IQn-~ǂa OK:5f \!-r*σ 3Ob?Z]$p}8>x_>1{!,`5<5Bo3mc9wn"D1.R 0kMѿyQy/-C>8i+4W;s:L(}tiFаꥨ+ Y\X;YL>I .'D1!rMuZ^QQZeT UEA0323CB/;#³a:źg t! H/%{!CkI(0vFkbJqQ[K½)D!R !d&FIg! ѕcDuŘb3 `/\dYažJM8 ftBOׂjt8W' CAP®i+OIאI_"z˼DY W0qB`åc'k ppܹQBj©WpU* ~YCR,Z85k2ai ɨ2&i‚HS)=89 g u2ȩF\#HnFb!iWc&yI409GåONKn])!L k!)X bHe 4WFF&,д[hެo "Cwcb6dp9j*F3x&RX”$38FŤ@>k4YB&,Fc:!8 &-QxQ)t I`vk! m&zQn1N,`e1›q:zD uNBSNjv5<%{Xc`ɖ 03/!)j (=e*!$(W a<`y7zΟݚއQuG䳻{ewaskaKrr~Y"D8|oF""A4Y+/@PƓLX(8sN Xw3tw{ n[fLivh ͐Dbug陰3v0%Dx 7 `dpFOwN/f &j 0ɝ &wݚ0`pvt] -d7u e6ϩᑔc)kRm9sY5[%mbcj-||3XVFV^}x RZq^B\+׹ % LMe_sB:UC0w6(| 0 AM Frw ϊ|!>&h|fTZX 94Ű| Slp J >FVDvRamV&L`P"dRH)2pI([MV(G Oޘpڭ*>~Vxk"G.6$i'Gs'Et? &G!.g4a@TڷDQW" LyN}Z tQSs6i@Qns`ڠX뮥Z箱:?zo@zzg'^BV$㸧41Y߉޺Zme5!*%34V>|,W#tvII'?s3=lCeC+}Ca釓b@(fu7Sh.xۻMn/ac֚V)vŽmR$n|໔-CTDwq􎋔њ07zHpFÎ @}F )UŞ = )𭈞bDoTdiWy ƶ.Rd1FrNX-kd8`rZD"PPY+3?m}>b?> \e`sǟ NLpI8=i:7lC~q΅'v\]PU" CͨxA|痎 T5umSsɮqeO݄]uOK -,揓vCl?L:4 jMQ[!n:./[T^ '.NTЄ+Vݴ*%,/馄9k+93J?*\ M#[l2Ef0m?+F??v4\8|lkk7`9-ɍOetUՀhDﭜ=ŵgןقN&@}u>hEȾtzIj3N&{hsf12R竍6{hϷE|:os9q=POiI>cxYs`Hcf9kDv̖;5c&8#)#;w.Ep',zH).=HubBTj"_eA55VRX! A/u^!@QGUPTZ9ħ̓"$26q&(W/ *x7b0IDo<t j!4WLyxdFN[MRS)Md@ 8pѼxzp/(Ao<hNgR6xNRaE#5c4w{ʨ\^q' <ъAG67Ln1 )sS\!֍6o!mOӡQaS!Ks9atMTG wKmt3\ClWy?|wGwyTb^\^Ve6ulGjBkVNqƬ4sd:.$ ށ2voN9klo MmҽxN,7yfp|z ]9Nyomb_C#&sck'vb :pa! l2;=~V3@DHjD"l6``t:(S6%˸t5vy:q2SL~_oZj!k_\v܃I$C6m]@Luy$gA*+RQ {ˀ'R& )uRepax]͢n)u.:)lDṅPt*ǣ}:h_ur!0͵F un mx)Gx:'V2rU.2J jECu[0r@;HY4+rU]ځ{~'mmܝPEZ9V('RM]}c; E 2myN V V퍷qM?ej#1{sqq$\~2j[Oc:>G| ֕/kn(mH`#Ea-Qn H1 hDd'a .~)&D+IJ{4x;>S֊.V?]kY/\HL/oϬ}Y1f@`U*GUׄɍCӱ'.™!_>T~7:@Ilm qBT Xअ(j'Kr۝#u84oz~1qVrso6F78X5}=5~'LJ1&xr֣}QqV2 L3T+aR™M$? zAJ" ^N)zpvb#+3EN,&BDE`LY%u@! Wc䔑BY\,/ 0:2y$v =Nv.qUbp;Ho&7` 9d hȜqPO$E 2tc*8CPjV(3&h  ]qO:~);)Y0y 8GBI*{OE^hfKǯm:sмѺeVB k+4H\6'YWD$c5āf˴;G9Rc&T ſZe-,h8SݵJ_^.EM+h*߀&"Y@k^(!4E7e'8߲ͧe$G:0Qo_-;h6DzQLnGt3y8(\ڙۻ¿!m MKׇwu7h?Ԅ #Em0V/pC.=)P^y6zc5!pzIZ^+D\:S6 !F g\Ϗm6!E@LxH,gىʊrU+Px 98RDEX0frD(UX1qBAٶ@ 4isrP}/&NU;O.AxOn&HI;h f+)b¼Dt V1F|@BvT6x$n=sSRP<%۬ H}%|P*dwy!lCۖ#(ъZCΘ)D*a i4a9`QVvQƥ8˺\ם)rgM)DWI1sȕ, v͜QSHs-, }!x \VX`Ԃ7#I^M}Ѓ!Ƀ *׼FV7YYU-]]qddddU5"Hsp>p0``o1o̚"H?9<ljtikWg҇}&\_-ܢc/oveÍ˻3?\=/?f 8Ҍœr13/=7 ]?>GCL ڃ7i`j;,1un\F iY:׺lY7v&n}8;]\,B-Y*a[%ѥ{OI+e%*!6vVYn>77'0Tay^ja j7Gۻχ ?1 dI=AU6sFca1qEX*vGMpof;BXian4C5RMORh.]Nۏa6}~wIrDf! )mVHzD)Ա@M>v;5 "7WJT=u5=twIeAro8YAi `Nx> &`0TL1]aFxD D̀"CLԕn>3&cj4-*V牀 5^$c&t. #ZR0phЛ:mP<k%DySbp-$ZYctg`Xf9Sbל$dsspf``뱻>7aqN@xŦ GzB2}qN8kYڏdp0uT o Ƞ* 2/ZNxbA]48PRPĥb@vb_n\48AFkEA2aEԑ2UR;F-'(*#T\nZ2&+6H3.ę18h0;^yIČrg&Pk'DG?0#300]4XP¶(x5paiCW"N8`\FNJrϙ5󷀒]'k\ ^ {-%[P Qd*_."QƬFZV®n~y) r2}|;iowcy` @Zrl @X1bnK/I4WIIVIZ:xH\|zt)F#q#h[)_ɠFǜ|!pԎ1t HJY34ȇꝾߚnV L9w.ӝ8(X*ԉ%4 WC=mY7A1>˃uHe*h-TVBCp)Eܶn R.26)Bb: R6o4 W:{+_>j[u7hI@+/ KE<3!7Ƌ E~ ʦT{2Bs14-)}qS7 ^˗˗S)oh)_u\!jGK8C>-u!3z "126gsEk!c#\m=BKbi*TB"+)R2)bm'` 1"a0^NGQoQ`ʕ&t)oO(V2}Qk Kjw2Le.g 27Vf͎֚؉-k֋Sn)f Jr>J^[~] = =7i"ˍU}gS8(_z;FP۹Cg 0!P,_{ȬL :*~p{59"WE("S'EA^l(E!kWBS+pUt/ #lK6C ׿7rzSM/1Q8|t9Fj$e5\9μίoȌ(5-;E[Q7C35֪V&Tfoc L/mZޔ]ژ6W]9:kdr}sXS~Ksv_PEa7Qq ͲOfT6әO p|W\QFa6|3cSK-Nx4[ˆ0&@a:[3W\=!̬L߱ u5ٟYΞ`TJW)Jp.F_%\븠! =tzviY9c!4!\/>eNa=9q$fql%jA8I؀N>КR G ov=cQ9V$ ʏa2,6ҞA×fK+ѵ%)Wr_0V;%Х/ڦxЇkI 9hδ%$[.bAQL 52NMl z>J/49!D$bH v紽kr/@ %FUf}n}9ˌ*4Q|-/I-c|yx/+bcj7DˑՔ\by9Ry9|ĎyKH&’y9&.))p?L29 S^qqy8`4)rP@x{JP"=E@O a5}xZLq?x{%)|,;?_YQgRG(~=A?7` Q(Xf_?¢suV~ѬqH…5ўII4;\zM8Qh@[&b&(%|B `?#5kF58,w׳3Cl8jqo?MۅzsAvEi\)W߁ouclgg+Vr3;6.`ťA*SOÜParnpI\K{!H6`#(cC9T0Wa*h+؁$(^x)`+N]jI臚77)m*MO@3f6"CS$j͖ne`]#s/S|nwʷ.w=;YS4 3ރfv爯x!2-c0 z}u.i{fm]4~s 3HK F_{.]QjD5M0f*F_Zn{=v:v=24$$Btay &0'>wa [8U'KojODּ= &FfWGzprGX+Wǿ㳅=^ }DI]Wcf^4קOyCc'w^uqQ]c\ Ds!! پߚ-EErO70G&aH(\$Ow S%bi M-]=6: QN:PJ̆Ӟt5{sdA(z**dIؠ~dagk8raEP Ƅ|בS'!>>ߖ?xБjqë_< -n^尀I^n$&$mw*fPM|J ovu[-Yc txD~RxI29)\ 3MyV#QHbLd$ ~6uQR;ISz8F RqD;aG΁Uc֖7 EN(獳TioyƏ2vZ 0…$~^e kf-Cx?Iؑ'e[Pws>z "7~;+p)onWއ:+3^׿937_\\ x>NpknKYaH<~{v޵>m#E5nwkG3x?TKNuyMrѭFR$75(DF@R5P_wn%o GK?~#w<S.L: '"aUӚ`{Y7 "H`MĨ9뤍%ʍWHĞm=~>x%ާy9jjԃѻ,, V>0 (R"CXYU(Hkl)!qhu<;O-nGU+@hz,xbnHv!zn%H&޺Eg-(1Z_5?\ze;?A9w*/\Dd*V7pۺ ˃"2턮yc-Т֭ E4Jd=-h-ED'e ڜu&uBBp) cC.yaH>˼cRxqGV>w>˕ ^o, 䊕^|ѮgJ*B^(l ;?̒Dڜ ƀE`| = o| 6/E[%rH(F!$RBqB1>qpWž(yQˏhu퇲)ϸxq@TTPP"Igo&|$ҏQq_*5PwR(TUD*\R.Beٜ_glAOB XCw!6yGY )'{b}x POrhh'"_'Q=JzbC4T顖_dOP'V)zbOXdǣr Q=ʻA2A{b5갔d-zC!iJ|ܹw $oX\RwK)7p/GPE5jq2̵8C;$XzP Ga)֎I˜aܥpIDrnd KrK!RN2*0!$UhTR/L!FI\d@<,ũb^#.Oj݄zzĥ{bS`F"l$ M0`5tpV:)C/|Ne[Q+H2AcyP$.e2AO73C fV!!_FDn<(":(cN}LsnфnUH.12rum8Tľu;ečYhBZ*$ ѣeJ%q3ݿ/Nb X U/P8js?w/vk.vk \t}>܋#9ݹF+v1U}bM}'O)}ʼn\}9Wsȑ<)wWs? _toL7c".w]| ]}ÈֹmqlzW h!L'~yx0{0AH$IN&d:Cq3郎|cx{</a;aƽ[L qOInfЙqg̝-^&p4:QMFó㙼~ ԊU^g=٧A]=2`6N=KJH~Ե20-ڮQL2v>'/> fj(,;ntDpNŇJ!IltP}g"(p)3K+;JSW1fiZN4Qq "?p֧Rxs^!xBcG Ur ڟ&fj 5ķAb0]+'Z8jBc!BR`Wov~>\Dʄu~כ9`-F7,VDWCc4J@*Й5Ͻ'ɏ1U8U-NH)"B 4e ू+XiVP".o=eAPS-31Gv->U|b' ܲ9ema|}|EP VТWi5IXJf_{WUr8繉 ƻ(+׻(AyO`T}8U)J"yU\IZeV|&a -9ifrJ&=o)Ab_}m,}5vɽs|/]re9rf+"GRne*gmb|ݖ6(뉔cW|Y6b,?W${;Z->)ϲ]CfGcs]Q;qw ٍm\ƣ(mxdϣs 6Jw-f4 R;#7sEBgEBg%tn\%rI;mꓠh~/p*{q Q&T@*R BuRy%J5hNRJRKkHw7aTTCRQ[ c(Ը\{S0Ps*xRUa8q RTM)l~R5jV^w}r̡HUΜFkeAij)^>SEZqtzk%ۊ,ʒ*p7R+m\-JqJD% 2A Ӝ5;ŋU~x |ڑ 8Ϗr]89 탻݅1n%b#P;~02 (pe-3<*G8.xT.6bT[d贇M>qiSF}L?MEuԜbWB8Z6qủGY%RJ-=<%)KTXgԡPs][l;oC7RڄJcrm"qFD8MKgziԃ"[0$J󥴘Gq_ᾟ̺SEΙ$p,*E."+( %YT#($s*])vxs:S}5K_9R= 6J;DdDY;Mw{!9dd*݂ HIRͤ;UFtТҥ[5LisHOt6PYr:/'-z{:!7aaA>{q /읝ٝiw/?:8$ LEXdi rI{!P81 7v}v,hl6noCO><tdwc4|pAcrwW;7t븾tfvܡ#r V${`/pvl00a|Ϯ#?$p>=x O +oap7e:?wso>]~|_?O{=r?2o<-$^c|H5zg?c{I}?p_to:f)ӏot m>߁{Ԏ=u.ta BAtw>9&=3Np%eR,̰4tZIDy腓L A@^/K&Y|>zZ?G]D6[K-"9w f\hBTY?ܛó{O?iDLz̼ˮT~dWC zΝy[h;vqfj>of׽v]n9s2Iwx/Ĺt-|r__v{Ap ,LǫW/zź!^_aNbzaf.i*e=m^R#4T5J Q*5>'Sq)D -j+e1UZ'CHOG@IB#Ivp)a0bDpj6cTH$\+yyk)0wIF5Qpޡgx܍ qjzEHu*ldC u=ҁ:HGަԣ'k% ]j@eja) !4TwiIU*b+SWg]d4Jy]v$5ŁTxU"͘q2^Bhf+4~3*WK/i؟9bYNED z5\EԤ>Mo ^'Lcf2̉SGX+'0 `8E4G|3ƫ64b y yl xg5KB/@ˋrOI!QTl-;KՈN¸i=gm(G@i.+*Ke0zI(zwf2 ag]ݾ7[8'qN*x{=en 6hoVWnQ,ّJ"Tݩ0StEE5H~)5k %-'G V^>㄀\VuSB`:D!QC}aT^h>GqAe:5V9곚bS*QF&4d)8R!x82Fs"R/`A!X:@覈uɏtO`*gXeA;kqHEnϷ|w ۻ1 $ :0d)dgj׈bψ"_UWWUWW!X^bRY+`s)&"w0u^!/IbBE  ~Ycvƿ!(,JI^ A+Ri/'Gkt9V2D몬J&uU>\dJQ˱ !̺=V)vYAwOa*VHѫJUA QKȑ z?rɪ㻜̢A`"pe&"OrYFaؒ)v Czs1xg+nf"Rq5駇)jI{73맔ψ ݂kzm {Aow7Y?3~/}f;+ьϳ2͖~`Z֪峐c]g)i38A6I~Ko+!jœ&YMO1 AC$)FnU],nAh ]u.EĔFAWe7I)GWe^rDlVײ <ײ4$SK6E ǒs޽ڒCP yBOB+H`t !S†BH,sbr*Z]k4+9>1 1Z- 9JqB۱َ޺NSTʛ;QDYR lܔ _ 87hHrb9qRQZV[3[8;Zugaw.T(Z}V-`Iwm1j%orݏƃM?ݜ<ǔ{ Y O1}tvBUŠTmt;h &*tCrS\wؕn.HamnK*ҞVyE[h%$tA֩FvFZnZT!_9V):ܑnZ"](ݪbPuQFo-TYZT!_9V)K] ٖn V"S2[@SM[nZT!_9SSy} k>+ ?`\OEc|=SѤ ! 0'ޥqb wXŬ7f]nvc/ -1rEml ѣY%jsD&׃N2b$vpg~z0:6KoӶ=Ν;v,ci\OZ+1F___\o܌W m&HA$qՁ #9J~zFfO?xf_?@ۏ6H9J`C~;hin P[FLJw_^l}_vXQP((9 };&3,Tq_+ SŸYH&i3If"yuoQY_^Lg<ʦt##@2BF!x?3E|P)cjy waH}7ͤŽ{$+hlr/Ʉ]'fnv6M+dk[lK3VP3'&KZ\jh/C$J`F{)]Ը I⺟VX]*׋9Q0~H:UɧM vǙIӴ"\H "U)I!"U)w_:HBm!R^z4uq E k&3󹽟P\FlmR<ߵt.>'3Tz xTA4iOdt[ɹs_;j )i2Ei3e+E&(1C}%JJ[Lh<XPe9XVz80>1[!H"*Px:Ln$٦6zy +v` \Q/ gmp;S<`ÏAO|CQ!^OcSd뜃cL3H aL*)Z`:ۙ!'>ιYy7*3_(fJUFvݛ4]B|L? ]3(jwقK2FFE#Fj;pu=D]]v;Z]j@b.O8]tr Z93ғ`$Y{X됄YXVۛk|_M=aGAc7&wAM>]dcg+ݻ^_s[~pɗݯݛxY_GD~Q`mD ߘNslYֽoIy aEc.w1MU_m?7tiχl/M"_yGdtMg{STeԍԣl%W%h]@K1hFKc|E(Z+ZѢxK3Yf /zwW~(Ύn.-xds}~S& nj%rjjU̐@(ʬ8$\ =e4a7-@bO  C.1W_8|Z N5 gB$'F2; CBU\8|D <V``Bhrb!0u'$##; Py! &PlPHJ3T Ľk e8,b~5{l7}vK~?qga41,r\Ǚ*I.aBP sR:$"LL2}wP`iœ(?[|ĥrA2l:sECC8pŌUfd[ zbp@X ^ Pv$RB-B ^* YP o-WGQ<R hJtjM5-k:paa`$R㧵5-v-N큲jَT,tC*k!AP0Ke.Y*TǕt`\48i˃qt 2 Њ\~d6fD/zw[ l̴73맄Y,CZg뇇M{wCzo ?B-o*#$[v&x+j'^.v7ԊpO>:aYȌU;ڂ,`.`2q[bGv^76UM\zgZVl2s-T&[;gm٦nnyN6h.LL`%|Oip|&fS`Wv 0sӨ.VPsi00Kw Y4u\lP mtrF>mFR;x⼧ND3%ǁ;j'i};ʾvL+/u$k4Ъ]})ۭtyE[5Q= ÿTihAع?ypq:kUKoo Gv꘍Ssރk-yKJ}~_Gp\:ǖc8kTZ }sg=E)-rѫS~23} 0}&P:i)OsJO->ҍ=մ/OVrN.[Ͽ:o}b .Cu瞠yIT=k! w^~K,L̲CS"y!6XkvlOO;>I畵̜x szC-SKkٕ<:Ak˻(U8cQiraD{EG E&xSɍGuc$n\Rc~ūѹ#$u ;ZG 8ՀA2 >!'ʆDqϧHed(B!|V9AVE^!Zj4MS,X'EP&ȁl-3sӬr;lo<. x JZaHs0EH-3(W$Id[&ҹh=+(uᓉ3. z&2E:ݲ iS7wW]R`!Q竁p)-BBY' Tvx|Hz9K Js-R^fSU3<䰬sFJ{&%}NN@sLcEK\WZ:tWoUxFӤC-q ʐZ1[̷KaMX!&b#m mTTwڠ S0 U09.Zi52q7K#)diozyZt>R02#vtW /QrHc[37=,<<4}UKeKI)Fǁ @J}i0(+!`hFO]OS 4vIM`CsjL_mS.Cg\o){@yu2=` o?1?7!˜QQ?,;+(-`R߇$y r̗EV,QHٱ_7(y48dl*$]!OuC]~2qn9[>nØ s J9_>-6X,TjөMP/}6r.-XL9P2$zxXPeUk-ø{ dr뭧9L@ҝM'Q?6>x";s"*_F i׎3hكDph a%ҚÃ;P"pz.<=b 9$KK6jA* #̊+@7,L?R2 0=tfxH) V#QM5Y!Fb;H4]V,:ϧRw,p'9sKu-i9Dkr}1A0HWO]ϧI!/ϒ`@}.;f%ʱ(L{-\hZPoe!mZ`pYdn]ebgw@^u!3 dZ;t}m5\f$AmJ08["z* ](XG#.fnё I:i22@1!{JK,|Eq eθk(U/AM6"m=?'A СJ.slr-CSB1Z)%l1 K}RUXVփ_= 'ؽXr!cN'̫KJdo; @k9DhDr*֭5 ;#5S4}(s D>+3|<Փ3Eq7>eQh MU-Bd P6bK;>)C$Ӟs E_j]t3ӇR5ݳLZp NgbNg iE7NyRC 0NY1s|8J~&qyxdu\(PN9zԲYةh֒ @8afUv>oJ Q>Jh')Ԁ_)eՄKezݴ^,߫:s^5A sK"u͈CnIhﹾvsg[o.F75nj*J3T,o[O1g{ X*-4#a+W8P!)ǽgN0qRXב9Of2Hl׀WE٘;.0+겘ר3b:J+mr^'wͱ`]Nd;*f ) ak*T0K}ٻb&#@)CV(). -9(_@5i1)I!,M d阃5%/gs-'2̊ 6"Jꎟ#-@fh|2RQK]$lSˉ?_#k}MX!*wwA]j#2An8ԧ$9 66-\? cǹGM<=NoeWʼ(X$# faiiQiyܢA/_;&j=O(2r 4rVN#ΔˡTWJn=(h;4f7i>bǥu|(<q0mf=!\=v~~ĝ)JyG崸> C8Id`8x!8R2p&{BފҌGNR)˧"U0%T%CEqq(JC IAangip 'O:z!NjBD+\=8Бl\Η_M ) 6rOt+Xigčo=t@OB`4=TJAQ/q8ͬt { -ԁ'Aq B-zPQ$nFe0hsS\a'jeux)7%Ǹ+ePwKUK5빘%2P7Dh@HN{܆VNV6٨=Gޢ%j']S!"RwnsϝSA &/4f]UhMS9BQ\>}5lH xH }%/v_$~ JV{_d`؂yŸM8 ac4TE؄i| )@gRD)ůz<+r=_Tc9fa.kK_Jfp~ٕD"4t6 !rv<Tؓa$L|ϔd0VW_:EHKf2ɶ9zVg. w( C<$(Sx^LlB)"أ <VsRIt D}*hD QKM,k:k}; iO hw;~t:R̶eԵ[hhiBWRxwO\ y;e.H//0guW"cy Sf^͞o1X (H=<sJ6|̥CJ'k:Ho%E;X/G~DA][uI׬5qlIk [;%J'mko /פKѓ,oMٱl0dQ-ha#6cBƿݪ;\u1Yz{DD^;rH,WڎɁuc.B9ϯ'Ԝ"LiEfQs {ʛG=1쁿ԦFk .ᘀBYKF[gXJAs4(8R vZ…2hZ|x0@:KRlsh9X`)R*N2yN95|,rgǙȇG=6Gp@4r}WlxfƑ_W05.yR v@K;{Iۡvj@K A/e@1)/%˕g,d@wP0GHDfTd5X:I`s\;`!U&H<zpHunxٻ'm$Vط[WOr8FJzPAY%cKݿF'8 Λ`މ01 |v.$^Jn'@SFc@nִhA 45,vwԴ 2Լlm畏O F21>>Nv8 EL_}u6ӫK# dס ߓ96C7~1@AڻnC3jR߸oq`݊ !KӁH "!sF(Vk4{54,'djI`IyU??LhN0=jB ŀtJܟ3Ņm^sHҥs7)1Z֨(iB '2R;6FPJ!ͤ˳pYrFf ^z1CM,mN~R,掩&e;~ū϶Tu6oT]z2T*C,(Gy JD ֛%zhyU4Obg$1ZqrZ?61uIP(5V`V'x8!m !^Gdxͳ P+x!x]9rd[1MiVrR 0"VI;y쟞QeW7c-Tz4,DG4ofLqVx;z v ^,μlt;4ڃnvHN[Ek"=[8#=:m@F;KEP\ +etbePP*EPI£_u9( r!)ޮZF[(i*/y74p:f\>-RJ{u:;GM}bƺ {2o a!h" {C fH(C[Dma$oWoVl*_2$KKN]n&/+ְb[ />5ݓ̞#{j}Eʉ~xz^\eRugAj>ԍ\MW?ŔU$vޏlQsY!v6"۷.+,MARo8;~F`I;-A'QH3$OB ,oݭkฬ؏ECjj t#93D~SڏPtR?$B?qZsuQ3V\Ƈ!|wh]Ǯaw<;Ŋhf2(Hq*@JDX`MVBDIwF,XeQ! l4Y |cF2* *E $@THB;go!9(5s(qeRJl'9l@1S2&6f%>y'r8IA1?Lh:Nq8s2Hg73ߖlf},ʞ*))t2K1|z9Y$|޾4 &фͶ^hHs©ara3n#OT:([uXb>VrJ"1]&"AM};PmdS(`_l?dg d(ì8>CC%fs G=9D=8 :LRhSV޿m}ʑ|{o.%(Zi4N9ˌLmtLg{a(ewrz#U֪4^اs޺r}9Z?Q{ὗ*n:&=qP-KToo(nsw1UPDk DB*4{u;t5eAv(ScHG4jT1QDbܺzR-f̮Cp!J[v:uL j7:Z~6;؍tLw!@Iox`OGGYUwߍeb~Vgd+(-0'O([ Q)z݈בtn;gDo퇼 B\-y 6 _ ocWqln>7ώ>w~iɆڂln]GeǭJ,G3K&^j{D'؇@`/?^6Lt/NC0_VڡЧh/v5 f;R!H'eɃM'#we=ދA6RU+xoVm(f z_nvtj}=LO$SAv3lgBX{tJs얊)<_0"³5SWv]HBEr"zRfr8;S$sA%~j_Lϋ|x̓|7 K/ٓ~3laơ1؟=M:\T$߹B6O)orЂDd _?'ӵi4w\ƈ?iҩ3[^chdpm>{rdX@L(I QR`mժAss/7a@| (ub :MnDS(TۏdzEnfϟ7\ NAMokn$vׂ2a??٧D~]g>9=Rα| {0rA &.Dpdd{7Ɠ(_Mo8YAdx^X`LGڃY]kbA"!Lr}~Ky]ܺs\});2NG9g2f њ$JRHSM!͘I|:1( w{l%l{èDJ1yq$ܰ7nH__}cĩ dPMHt4e2!d 1ERn)fS!)O aV$ sNNqxk3GAk`I}Y-!Q/H\ fH rLp, L|txߚr J0\ IQ` Ko-XΝ*+sF?I S䑠HO,WH!X1RAnM) Â1q]:R9d!J|-EFN$%L`MVZ$-d^~M6]Η̈́)ebMM nф+X &8hAivXlgHb✊d_e"Z (&4oۥ 4JxD-ƀPL@FBfAHi&3i6N8dfka!k!ӌСrCN$(WHA{T,ay2[inݤNu6u3 ’3G<DLH ?$a')H[pFh0v#%0L\]c@B5#1k r/о8 ,svɰ%5R*ɠu8KQDIV" 1;a@SdB 1J pVΌJ\( T bi) !UA&ːa'=0N*N@$U7C,hGDV>̛`ˆ{s}׆t)yY5f$+Vx%\Nbbᣩs ) O,fB &Hf,!FwLxH\gFT hN2dȳ@ɟh ٷժC8}zR0y0J~w(dQCDCbm8!*\f $!$8J {-k͢`ٗinQc >f?/|)KY$c Z%bv"9Dcs)WtP6I'DI*T' K T' z:I<^+OcL}yJKD*q?HJpt_CkGqhI#ZA[󩗘ϨLFgFeizJdG [a0in pBr2*3VW$yP@5 Euuwkh Sn}/۽2sD Kz%#<~[ ӈ'UkNX cCLKw(]W(tS 4lRPq{VES$;)l(KpP(lh5AXH,&p$T`槍*r*eNr#o (*'R 4X;% TJ[!Kv͆bi&"OA"RbP \;q%0pĚ 8zɴuU 9} ]hbARk:hv0rcwk!CPo,/L θh aRR (A;* 3Z>9J_1nYuU,RhzF(aٌ +kcF25kR-VYuN]/:P/oXZp QO^ti,WRjҨˬP-au!t+o&f\;ݭc)Wmf_^&WN6M_)oF "BV˷\m:w;@eS3L>6PIͷgG2A52"A\Rx&b癰BDn BzvVD ,HfB(ōѵ <4D Tf5{>!X!VHԶW6%&N+04#b{&>(1!ErP@5vc7 Li(WMb+T tC ںU N'g|j8j*Lp4> 7؁3wS?8Sk!2HK}? zӚzt剡~fKTw0C)#BkF>͢Q*\x5Is*E﫰M {UX~`ga= SJkXިrhBDl~nB=[ rLt&_ltZ453nmXȞhMn?wge~Fv/Kh}#Bn 6,dMȦagAͩLȎRm McqĊNt7lXFjͿ=TD/k4L0yDDa C1̭dVsiy!H8J4 # PtdHY#fRZH93 :Gl ~g&`g^xm5y \  ժ0ǠT9t&AgtZ?uY~(3lBDlj\= z!ѭT9S:Ft/I:u3nmXȞhMZ?]8Wэs V*) \[1ֆ칉ؔ@(9i!=#N ϰӘ +2.f#k^Q)1bH90 b ʣT`YTNP#CD @BcTHvBWb'6Je.c4=zTmkW^ a!{nE6E{tävE c4]t_x{RFs-XȞhMݏ8WMR V*) o53nmXȞhM)6DҲBS-Dx3AOŽ"+,`1'N$i领qpJ"'1JvWA+F&Ife!fCQ[[p] z-xjv"w_*,p)oփN]WM-XȞhMyOn^Gt+A锎z@z{bFsѭ s-)JGk=} +U8Fl-^fl &ZdSl,@q񵍩X"a%CJQ(ZP R4,0( R<R8̊OT-( LWo߿(0jj3};qR(}*9 v4lgNiUiJjMD6SɖYG%LҞH/T/}qrZ/aETIoUZd zpQX>]nja[gPٻZڄ1XJi8=8nsj9-mW^W^:S#ksp;n p,t򭨄,i8uÄ1O뚡ݣTlQԊ's+z?bzdONAJR91[8EWmi)Zm!`vNC< nOmʸaD]\?wMT0&w~^Cޡ. $nn 8`D5RCo_WܣrêӍOtMΗ kv'wtst6b:9|7''gaaIK }\30&'W֯?q\x HDO!'/Џ/48冘JP;&ÕSY$Wi<#jh)B l6*}HMIqiOBYvse]-ϣsR4t ל0J7%rP 7CTj:u-Z D`ս`3" #ZuH6uGx4M {-I8_aҍ%4Nab0qPI#gdj-kV425cAz?J[h%CE4$;ʐ1(NYSU z`EHQ6uv7?)GlsmԲ?6p!jvg $+7F U"]{uLwy1L/>T'\ؾX"_ok<qm6|`P6y_c> m0Oi'm;9Ohyށzidwi-sbOj 8_+-yghʎSvs Sqm7`@ -yϧ¾B/wU^v c|]3O *t $_wh/{־1^%G)nՔoqwji*ɨ720iDLPdBR9 j!z82CjA@RgL=~H/-+/!Ta!]h\Hsm&F>wr eLmz& x/Gv~rZĶ+ ] =oNs@;1G6֫!8(4J=\3Uvki؜qLv0C@]L>YZUkQ.R={IY-b;qa:GYj?>Qʔz/-8L[x]9{ql:',H)Ҏ'lEv=ԁv^8PUS+w㐢z iI'cf:z<\Z PlT 1jh!C?pcq}ӰfZ9~ش_%(_??Y*Z3Q۲?bqv >cC }2vzv) uCD]G })?YE3.5҉XJ$fulyok}55O{=;PZ`$Nz >JʵЎy)LQr+!4F$t,;A`F%R՚Ȥ.6\^f՘YWP,]gG1DtXG-J}d:/t|~rw3ksѷLC*)q|085KZ3SD>N*&Ѓd)/Q'9ǘH8yN7?s!Ifµ L}_7JpR{ܣfbMG*(C +†!” d 6H: k"1Y@EJCÎ=ęC8ャ|zӵ5+'J~9#~z Sy.e8hˎ 9@8SrFo#:hf3BrZnPFހKl$1cJۘ1ؑV S[n>2!TmmnC)ZxPdV) d L&-$~Wޚ#ʄfmۺg|ˣ+Y->by{G-Tֿ1RD疇A$(GOnpI!(k`y@Ԅ? J$|Ji*[#uhP m)v)ұXsQݔ?nf?]P4j׉eTrI;[&:pP KtM%gFou `|hL|,'~4,ygO\1~KK'#nmOfk<-l}Fc r| l5-ghfXĮT;Q7F`h CJya v2}XaxXɯG?|^noqE߁6p?ܼ;bߢI̼[?Hhm,~9[Q)UoZiFpi $&!yQDfS-x 'A0+YU0d]esځ+0l*vp(IB-9$4|{4w(<8gVq4֕t?1f*E>p.a%#bz # Qfs9d>h fQЃ*AƲ$ cxAm(OU1Nfaioz2.q[fHSo?CPZ-*CMD!v}ZYOBl7`FpZߔvqJ'qjЋc\Zч|@X>d( C|XnrK Pd*w!,3/(q&cw<?f΄՞{ 0=mW#e]?4Wݜe>A՘iQcoˤxJ;ei鸒l߶z7=:O'v0]/]͂qobD12"H'q=8^0tҒ80^`>wv8r3l1Q BP]xAzDY ~jM%GC+U+4s>Rsb)%c"!]8 +Hi-%$xS5_)[yhG<D^W|Zf.N  NTZsK[{gC.y=Y+~y 6 WĩqwZߍG8&xqi2۵A4)aزFz*f5qvݎoW4Xb.-Un=>&K>ggDpbɻVsϥ @+-~,"כg4/j璬׳wEp#l2~2uD21 ?E@ \#}|/[݇oVHRK?n&U鮨8|36D=,Fsh7ۆb)NcNXr"#us<"hw7_v:oܦJYJ15B@ȔCjh ҀL8Jlk)aDIv8wDc#'NcH%USTlj;d"rfH| [GܙXq&sNs{CNtJAL*[o+Ҧm%URJRN*iTJ!KJFFCT{צsWĮ1C[70X%f۳@ 5NZPIuL`H82r܁mYI%ȋjC&곜<K +RP[IE> C%i%AT\K4:/h Akl%p"NQ_'`S9Kta_5Gx @'`uM5TIN@t8I W(r[ ȩOrl똮V0& Jq5w#,hgԲ㺛-jvL"hAL3.gRW젦̑f늇z^M%Jݺ`詖8838. { q2ÏAa:VeqT|6gW Y`q]4F^tvR닱=T!b8:u!\F8S٢;^3T\ M&a},AJVt.T a7J?ַޒ[Ϥ?x S4J %v'5vR_HDӚt>OՆP%^4 -&O(w_bQɜpTyǿ1.j7m:֫RCN8pz.ǷwPm zejۏ-*;x%g+S^mLi=*ŸFDM]ƠmdG+B\-3&p0=o ᭡joHNЂ7,_i_SO .agH{%I bo ۀ)yK`\X.qC9c]pA(h!4 f)lLCԮ%Lإ N#f}Ԋ*N^fRr^gջ6r~?ږbb..Ofp/"W,/I+n0-6wMJCk :U~X Z~[Ňs*?f~E%o+ tFR}ܦJ]38}*7YoS@e|\$ .Ɣ@nnK_/ uΗ`;7^8l):S i?JT64hRvݸq9١UwbмFt7mDQC%R0n&q<+lOn2>BߔIobe|3gni#/ӫ$ރzebOS)9X\3+r>_FiD\}B5Ȩrtkq̨2g'YI; O j1b)zvY:D=3\usnJ}~L~E~DajٻƍlWdr틁~7 g&/ ¶nɱ^SlS{Qd 0ZוּO bi׏L񻘏cO.ij$X'ZU D#໢8G.@TB-]h~sH2] OjϮ=v漊)&*3L1,}nm3,Q.~ג0}>Gd/ׄϹS& ƃyf,ΝG\fDVAD x>SDޮ#LH4"نbdJhCzE %C Mp AfsЇRMmKnkS0-%GkjK")irA88Ξ~.fNϜ__4V *2e竵+aFsR;P"?nF[*u2!$ƺg*gr'Rbxr?qkyex./58O5&G 0DwXmw LnTށ;h`/CjmS9ÕW[b#90aVIIޓ}jمBo03RyJ),,1$7 㐰 (X: /y?ca vHιEء %Rnm~OܐsRQ &DCИ+g msb ۝Pv9qy`!dQsى;~noɻw0S[ byd0XfA\Kn-cҎ9L: zBnu@\+ H,*9B¤DKxa,D:icCLH<"% H &EsK"Cn]rm9=.Y7:Ѩ,]EyV+LJ-M!c8A`/!KEE&Dv\HD\!tWg --KTwZD)šR<~ZH,z#b8ﷷÛj͚,Ҕ=u ?7(YꊘN-2J -Qy[(茀ՂTRM{ L8TzRi+b.֢TEjjc 9AkV7jRzS]9r [(Q2^!ÈΐW0D7-Tatt\:"M'Tt5L` )a+ߔ~ XM!q4M47)v%nF%;P}Ҫ{N˰(/K(fO J:7dׁ)=㥩d2Tt; P(}i#c;n΂y8}YҘį>z)t Lc~ݎLfǣ fir~6Cu턃1g>iTY*S^li(=krý7OoM ^6kpU:Yx~ۍūJ8|3Ĥtz'.c০{ 5O0OVwoFrt:!zG,>tG;:3dxϡaa~nRRS@"5Ȝf6984٦c{ Ng$_3eѬDz>C̰Jn.~t7 LF^I(n&^"Ka`\.KgkZ zK |jrBӧK .5 ûLf,6Z5Rw8\6kZG>|z4q(w_9ҥZ;^Q:XAxo߿|oIc3흷 V֯F/ЍxއnލƋ1oFe0"El4nY7P ! ;.}:!L>|yv.>F}wBf$5a֟wYĭ'&03m[.Id[Nn8*.7ǏwC&=mN}: Tq9\@7J􃿕ŝ~:wwg9zUZ}'%h]Öm̖ f #];|uz*e_3}x8u]0F=]Xy>spqig ,1t@}\dֿfѥAF1mbE]'L|ﳢ[n>:4̥$4y,ętnu:cO4˨^M'.-`Zp [!ߏ/>U,7ޥq)͉663f9J6JFqYkt) UtMXr$:EVbr(Vvrt^4\]ys=`3DI0g CV3Jl< 22=ӱih5l:)YW%IRw>6~$:xLχI81(EJ 8_ NqP dT gI :I*rVl)T'AtlE6ssC湠l:F@)MRI,ع.Jlpv.pY=p "R,eӫn H,S996{`a1M "a6~:B8=v^j%_LO,/V'BaKfۀvfysLJ$ f d쓎3:;?u]>]  >3{3x6M׿+4e?NoG?{;")_e @2rOpac.6`aCg嚻YMFJƝЈ`0{ Sj*Fc7#ШAݎoga:` l=n罁P|59RFhH ,ݷw5Px4 Ei pBWK0xo Qd$(5Lgg %F|.1߾$Chμd)qϼpyRIȹ3ٹA &7}Ub9_<7~K=[P䔣cJx.DSJ3f` Eϣٷ`^:;š/cdr”--]Sb+UR\+ _Qxu>kcW d D cxR1l ̐xiG")aX)s(y1uD]1J v߬Z ][x~:;EAS`o;"WœWmf Re0Z'>ܾgQfL/#벇'^ڛȊ[W׿\i|ht@a$Jw JF>s:"e)y\VX5TFg6ʼ/QݙK佝iF|LL)9-uxBITj o">x DG +ɹGPSCδŌٻ6dW|9nFE,8>TxQ]俟!%D;$'؁%^fFhaTbm#XB/e#(o" IT yP(5Z̹"pvcXg

խ-P5Y;cRe8&K [Nl꽽 !pm*Y۪%iFhmUbχIaz7tajLkz떍(|9M#_'&-dcfp?/6\ XJt.'Kl?[" >T+a>먹LVloüڱ#7⵼^fy7AxL N=6n%jvݲkuM[ r&eS|aͻS>Aɽ&mɣS%v [r&cSOH{7Q{0[.ML'x%N[~ޭ 9rͱ)).ּwŠcﶿbT[ޭ 9rͱ)8׽'[.ML'xN9ݙw3лa!GnY6%oXnȻeb(yn:%;̻eowkBDl0(r1hb:Ļq,d[ޭ 9rͲ)~owQ3Šc&=[n'z6,M4Ǧ:=Mbtrﱉwrv]mڰ#7,u@9Գ R&} lvu~t3A0ѵ e[ njc6gu)H͚<+߇ĦZZ4lmbSBvy Z~Oՙ\C:b^~/H3AO-wx F}Vg{˄ⲯ15Vg{5Lk}͙ ,Ycט[ }֥ 5L຃5fD_ckLgYVQԘ{5fEe_ckm}֘!;kGQc-WcVט[ Z՘טۜ ݽf?ט۝ {5f-wxf1EY+;Wcטs3l՘!hטs3Pչ3GL5L{5fD}1:"1sw+1CcG1sLk}ՙPؕ3ǜט[ ՘am}j`SZv !}184%UO}kܭ8Mn|pzg)Xxv=f1^-)gy) ZRe$!林Z,%Vks^}sU o_^~y]}q5A`isJ H]D,u%rRzpS6X%U}Rvv^Xu4zc*Y`+qr:|r `1ܧK*ؽLn@wal^|ݸL/2q1HY_G#>X^I?ŽyTCr~!~D3~۩C(aPS A$p]HAM$H1M0S`jAgAq/K L >WHX`b/X0ǬR!Q`.BQd$b0WJaS"60)"LccR9j}&2`a1lLH#Ҧ@|_,< {<,-QhyBh}U@(x`RƘTD Z72> aa֞YBPCD!742(_y50ɝs!7h_ )20?Y+:x"ld n@$'RC~ BBu5b}p洡` !m .HB̐znp)1LgA(z8Ć`f|9Zu[\7\ D-'xzNggqwg<YedW#3cj>5P0JYW6=+0<&X}'m||r]_̹J߿$jUxॷ|Y#WA|4]FoYs>M3Yb<''[eKQ㷋w=DhF73wׯ_kR8',rWU2(Sz*R-!1#@gV:b RR kPex$cXF[9ZC9,zY&DE)aq&@@TZ0!|a/פ6ԜUSA˔-fsR2>E1ex{ g *A` j)Fx̾urco*Ƴ ~,'1hf4O~$ ʉ/^}jR>WMC7J j |7!aRqN (f2.gZ?U [B=xʔ9j QfvCMbb~s|~{vz 4|V)fTmUfx`q1K)0PJ;s@樶^ҴkwA>n0ThPȖqN8V2juQ(F$k{0]O`y ֐l2z.~-{ LZ8K&OVi**!45 nq$\s .(CGiaa]xB ' qgs'h݇2ץ,D7J8BQ(+ɰ~h=老0vHqӞb8}/Y+ৌ! ^;GU)h33vΌS&,9Mk SodjHU٭GDH\va'O|fp'ik!6؜^|e7Ni1 VMn <( oI ~^"0wɽM'#ʗyҜq2Z܄1w| 5~] V{&Nps;WTےFKƙ-3 D2˚ryϫif e%?/PMn{ +ڝb+9c1MCHI8f=zկݩB gE{>UN'G{/` ;k]-,DzN5i꺪EДi,nIݣ$oY{ iArŵh_`g5)BpՍDRQiB5歜hbϣri)9A pZ`##p!L8hI |.)[,DاFE6d hm3_R\PDTt!T "+)8J q@HznkpէΊ)|hF_]%/gbb2 Z|`+Z^MĿ6$5WKIv>0FDf+ p>  Ҁ$(4Hذv{~jɅ?J ۷Bk}w^^I~#|WS;TP7M0Z?kWr0O Ϗ42}( zP-Qs-{M44Y>7e(6g~&vBƌU Sdd˳\.?)vʤ=i')(Ŏ({Jy,*YU"WpᤔQjߍ]9\e,$%p_NLiUY[TKa:X1 dU;=;; swZz77o?5UDVjMJ BI@"!BYA29FSOwaM`1FP>MiBp+]dCG 3*ӓabO4G˿X%Xj- *²ItbKb4B1I<~$>)#X5a(b(Ckgj_aW.^o[p*ٕ/N0@DmHk;~!) y!%k$ h'ו~ mf7 yAɁzG<=qcqZϝ0͒Nl|A'Fq}.Vq׋@, . Z!K 4h\ *<-GA[4~;\կlLԺvݖ] C)H-Ϟȍ\$x[aq;AOpZ3BĖ_J±ZwڿBd^EAHGfYq852ҥm)H+6r_{N/dt4wN=|wxs;[}~7Y߷?ن[[](태\'s}BSJΫ5sM 5WFgPa+văi^DS}gv  B\ޱ w ^gFmA EϕXze~hAL0ۉSmXQ6@SEm%?QvZ9IMY0wm>ۨ/+\ݔu$!qm%S\4Ub":mhhwI*n$j.$7.d<,=vkA贝v;; J΁Ck7|k ~ŏtOσ~zՆ=y%԰aKVLaLf>)$as3\@ gS.&VaI՜'2I7[CrUwX\u 5,2vs?gU.CL>Wo*[gշiFag(aS}9+;&3of-hg\]^|%D~XcrJ2;iEx2FDKpVnS6BRT+%j:t>q+~9? eإ(c/')S* +t/,R37KVDobAa@©\1EB"㥒&\H;{TQA+R)xb O-)刟+,A(F7&xf267 væ 1Y^ʥpMozy7QH(CwZH[XR6r>wS4X9GY"Mo0LhʀPެŤvwL@aF4MPSMIGVՋϗ'T==ބXt}l^{xH=s@UP.hwY" F2O<WդLhH#Kǖˉ%1`iٗ"A*mzsy¨sv%=:FAUO<UIjߨpqoSأOf:~O`@vFTÓ\OI Kt;oʹq6SjENh$ưK`(ΊЛȘhB wJj@\~V z*{e;AVghm0n[_^`?{ Kc mAxʹep3Yj c\0@&{U!)#BV2ލA. gR+qTXHWh[q]dNP֤px76@q vck4wň%bȦrC؍\kJ (1!9%V!v*:Crl-+6{%J.Mݚ('*hpN[MjMRԻ5RK ;r')SJ>(/ÙxֲQb[3&H!钧0'a~g"3.t{Ӯو\܊CZFU4z6ŭfv҆˳ԚESvX~k DBBg hE/ uX╇=6hA4GC\87Rona#Om4A X)3rG0.پ_Jb=n]Q?̖lv|lOtYUy.u!F?C5hFrem9vvhT2]zk|j[-h߿E9Ә=+HߘT{4F(d>緉?v`T_6nÝ>2-EйgF̀ibݡ}l~xX,XeA+BF!/-LL``9˶ݽ[\^N:iv&Fnh#Eh%iO\Ma@EAe'Mu|pl}2@%/E@ c 5aǻ=vJxb˻'CwCe'f>Iso7sVkT*JNng7OBT ߷or"YWeӿ^{<k}æ~~bAMnbCnz2L%505y\3Drr:yӳT-[]ΆD=zho ̀%Nҧ/FjRc5eMېK }|wEOقdMd=W!?E! mEuݙ2BW6;FN7cDf3zG :Ts 'FwDRa30<.s3S䎁YPaVʔIj-J7N?c(pu2aIZ0j\Q:AcjvU cuاQ?u|*cGkKKQ4^w0O ,;}yݖ7}E}]+VcU!Ii5=0jvd6wpO)⠅r?AsPNKyefBép&@4 ==(FϦvg$E=B*}7vZ8!e6_- }c|+v-G'ͬLȀOhez?T3 W)r̡aьuBe5?;^ERv}df+;/5fqِ=ux!lVB2mFh{#ovA$>3jF;sՁ_>K<6+8!H3HsR%~ӹb*0%$@ 01F%RH_mxI譶o8!Ϩ~MxϬG39x,7烠ZўNыkƥ]Ԁ"s~KC I^L #ͧabk K)D_!- :,E(3Z~Կj93{#G ̪?ٺQ߿ſbWdRɛ8i\tqw[/Q#? <湎6uy6zb~1K pj<: DpS(HЊhwqo5Pm-}[ 2>l;{G+ėaŲs+~]~ܺ֓,ȌqVqZ7p<*۽S,ލmA æy@KA7(WN蜘8/U:f)ʑcY$3'V@ MU@n?;ن/yZG_"B|p!g +5$8{>g^<y|Hr}җ]tgɥc>[5,nZr윍 {Mh^GjV!R@P5pOl3yJT(H`3Y/ n6N8642([H1/&_]A{۫L3'/ەeGщ2MTEy z9X0n~IVC8mN1(YCcwz*޵5q#뿢Vv(4plVlNRKR* D[1M3CZl nt7F7\h@ BpQNNq_sP;]"'p(m-|pwjr C*ߊ ڤq0R\IF5T[5 oS30oŴQظ2IbJ K@mPD vqC*QT IgหwNJj-u%ji4esJ#pΩ (zV!&.@g_Oh'&eCN]o!1&jKqF0!\l!O}쓔RXm؈&LQhDH5TRnFGh7SJfluFD $' ?~v{e~.åyjx2-ޗ7sf/^ɄPHRd$(\R*FnL)jTj{JZݴuV5oHzеVh7v0,@lq h{`0f-@FsR݀)fn 7q*4.ezd=Y[n@?B靱7Qr9R6_ucGI@Z-G]uoy7_35lecK^H&.k]=[' X%r᏿ӷo_whKф(%@TAk1HQ.ɧ: O?tb Tz0.K8\3 Fcx &%.٭ *-ꅌSftoԑ-">.vmѳ%1붅c~]WIRT2}l;ٳis97ȣ2^.tv.'i07{jj.~1u ۏfDy=LM0j_BuNގ?lγH?f~2[XԂOTJI3:fuУ=[8> mi oiuaJnӃKq3OIBDaDp,dya*?/xDy|SVupз>lb8/y4s0g7A5问fq[) Xց|gD5ɂtx ZMJn}7B.`,b R 4V[.<0fḺWQ /)h#h$a(:޵b+NL6)qc_z?ߵI؛jFr?HQ%z8T@yLo WL9T.YBgP]G/`TqF/@ `5TWbUZ"92RqVJ͉<TK9V 9_ZFI\Վ2W67c~Ol'2SV1SEdzyg;[B\GK6Zlm1Y/~Z.{:;Y/$9-]Tы͗={??*Ōe{i=I0s lbGQ;Ґ\Et$kM4`r1HQ1XFRҕK}ZlFS[hN8׭[.);FvUT`nٌZ.4+WѽuHۑYm=X}nU:J, 'swWܖ;P0Ԃ^l%:^i"T6wZPndǖEEv[ka t{X#JN3S)AJY\}7fͨ$[|_iڒч6ưw'/<޵e-קg<[ }OrngDcsn&hQ˛0gn6.tR\N jbv3ݲ|ZFH]4"q b\^^ſf@L7Ky?kd7?rbPk=Rɛ~0v:|ws]Gcb],hr5 ,l jyHÉKjsC'RaoʣPj窆% &donM\NۻxMkp;U|go;2&1i;r, !c]zKOmQu4Y"$^Q~T&\2}lB0B(sj'2ONW*+ IcB2} NaC~~2e&=7y^QF&1!EM:L2M:,HrrN&8U̙zcUe^35}]5QFj R2/cUJT*!m+\uZ|`sJKhi_b$e n a!m+ x=]fR\HZTZ,4J)PP\mjGJ\ӊ+TL%de)-+F Wp_,Q؈25D*e+&bFxŤ$[Ӓm6\Lj A54˷j:<D>Xrf,eNJ0 O/2 Wq'+|%z"3(4.2N='+ѭ$EUWW,mriq q:_?sI31&_!tBA.w9GN[Anm 2$X&,q!fgگtU҅GR3_ ؀kZtKd@t4?XIb@`I-U)p{?blLl<5H=P)X2TKyeI4/uf9p }w9I28sgj8p}ٜ WΟ^_`UW!UQtF !0  L(k7,Up+T^"5S}y/[t<_u7qbD y2)gA'9a%)I_3%LKf&{ހk,޺ދR<9%9|JTXNUm# ~^JQٸyד?t'<lqv,9,8Z/۞);W Rޙ^9H-ٝyz lvUEVP%_4YS[2v*x`xTINC ӆ˩ġh*r,{-ʄ&XPH'Pq:fxI#;^.9kxJ y„@˯ƄfcVTŢ$g Nnd2U,NfJfp9Og)~i,-1#ڐF+X[U5Y\J[/.S%=2UcnP=0UK5ʯC\p+Dx͢!5sU~O.u"E`)/ƶqãq aBnKst%*) f2:(NcI&vfi`odKP$IvW8͆@޽¢YgZ2tyA7Gљ +._yfٖ0DHUx7Z&_Oyd9PaѝRK)IN,4#DťPI$A4 bLSr.+9䃔#8JPeY̽W8B@'\쥆[r3%Svx6Baqr2E)HfUvTm`UՄKlFaD0aQvp9}>cA׳qH7C=3a/YgdDagu>FruUOә>8ƓM([,j1Zɪ }7tu)WPAV|d7; "0ݶhw8*5=M @:иu(`8|෤{L|GlN;ÿ.Hof[9w|ʙ]|)tB'M]nY#*|IF)'~2ϺЮzֹ[h3J,$ѽՓ1a>z#lLb;*f0I#giZf8HCma NB(L=:Ar6b}n?Ш1}P@.ޜZ71SRyo{BA2F=CyZR}S/oQ&Wv;nb4{յ^fX2FC/W͎('f@FT}-2 BTǪ"hc>]lGdBr;%X`BwJqpd}eio]gVƁ--:Ter ?2| ?ٛL0Ə}w4IQG+$ĥeb(9!- C8D,lOKf5l8>)%\>ۗKVRsnzީ}q`}N @>@#B@=P?="*ET拨Menpg4#(UɭBKqTCVJd`[){=wH`$rϚ\#>&3b%TPM(`pAŖZU:"@p?FSv}ԔߌZV(Aߜ>WtQi'mJ=)A)5lɸ@kFeΛJ 9ե,=z2.mQ"V#ۋx6z>]{tYz$^yɘ[<}RW]Z|S-_(JX踩\NQ!4oz]mo#7+?^$.  ޗ ȖW3YbKZovZ-[aĖ[lab?4C"i/a PT syx2Kyd_S|ӗ{w6T|!\C;.*?ӚS&P5l92Th> ?>cSyTۛՔA8ady1c?os6fB&KV >M&_dUqK s۹<07RJL7H䈂+M 榉@*OEx$@Bc\`M|Ն 鼋0Li^IjH>P))\O%3{;oq(^`ѧ/&_[s۟ݣ?yUUVV\ C Y .?co8}x^_|%eOh\8 7'O~3'>x?Q*(|nl/᝝_oQk= sS< Ysd^ځ/$ruGQAh K)Қ}J9i<.~ՓjYϘ@~vj[tXQe .za&6eLtYI wNZuSH9s}̂wL6֜.V՚=w:.~ as0G)ӹ̼TY9f\fs{1J!_3wܵY!ͫ=޻0ޝxAvx,>cC֪\Α]>}F,>wK oIiU&9h Uq70B0&Also}#Ё6fI6x8 1<]pk^I<nE,.|V|JOvgSʨ#)]`);237u_|2ynL?z[?ٰ,xKz-d!@ ͚=WbMSWi3,5jOBAޛW9>$"&/ }pљnnZT[][\Y /7E,!KtU'/ Vje5Zc)ퟌ&m/ y<5M I*f{Un;lujshhɺzZ#J8G6q#O\xljX|@m;Qui6'̓̍uAɤMhώw| < ѡy>Ǩ3ĝ?Ke>sg.l|ʶyL(ݭ8Hq>Lasy-Ja=XJՁ=,Wt ]9 <Hv;a pBnޙS;yv[/\IwUTo5M_j)]qMzj.Pj#w?~Θ:im(4‘B3#W$$`J&rp$Kd A  DkC7I>.E%iHN1u>tM.BKy7P6pI[;r_pPWZ̶^Q0Е-|V@_{FsbX̷=Bz ߩs|2SJna1B) W=M 65UQcֵ|Ǧ>{#ͽ3{=gl[HԫߞSZ)o֜5Wc+|B=A`Jb8%V " + e&1pf5 9E̔vV()$a1<",&L R\L@҄:H&µ Vzfh8KըIHύWHQ>dGavH@Iom<$ UJx 12W#05⃗LtKYY|7I*6EU;ߵ\k|@ J/Vp}!g4:H٥ j3ΡH>~3 񚴁0s bbpݹѩUGTey5R(mZ5Tm!5єNm6hb}}Wt+C]'[b9ɤ4SF&L?AI@sI C_BrHsfKUQr2dUR$gI*{\FNNLcnMccoג$+)( E ++ ֞F/mpI[sVrLkef~8SdBdHHsU2X4"d#U 0AfQr {j)ή^TѣIS1l2LY5>~a6x%'KJriO.(-#Z(0 .&嘤 C6( Mx\%f1#"@DAk < jreu8iI-1c1cAq K yVR*0I0.*Kt *$i,=X~"N |n&fÅ;qѢ fRJQ9U.Q)cE+Ldn:@4\Ka59A*4:Glx/s5|:-m ԙsAz&Ң{ OuWUJi$wG).b=O.7?+!^ ?~_Oo_@^h1 zLF? ӕs v1YFߔjzi<Õ:q9;(7}?aM2;\~z;H28G,gٸu{MJB.Qh^Tձ 8[qt|m;IBݷۡjG9/ 䖷/@8ft׮eGF-R2_/ڀՂp&RWeiN hR/*}YmKISu籼7Gh|~<7mT޿)5 eF䅘d2H4H59  :jfXhxp4ߣIt*I{ eś'Rq$+[n|>ƠşsZ-`ȫ?oR}Lݵ|8ekgc}_SL$TF--8c%$WJy*Lo(J-Q\:Jh8B'PZ>_oeI,%]~lA&Q$o'8}@O,T~>~/قf}7G|rJhҾdpVBohe8+9EN85̀IB G%3 FkxBr:%"ULH B;'R҉`M Fo]"IHv Ғ傐$pxb ZA42BM\$ -5T`ۜi䯴4@nٔ}K,`x-'|=F0jI[ FK78ٞp3Fŧ^Z7kXW(@6h; l!pm0ͣ3ti0[s2_㩸(]J!dQ~Ks8f)d(@J0R׾ٜV%-mP-O.ț_2nZer8XRZS=9qL3{DyQދp}1~𳌆L7_6w.L;`rܥ|𦜛4#FfwyjV%yo;̸с/lz[+`9O`ENO"sTD樈Qt;deC!=s\"v\ Ys@KΑ>jweG(Vv1cL 9͔a1H/2$IKLɂsuJR(MXtBVX_&Ww["|3g!~Y>ި^*Y 3{Ր1 OREȒ)hL/FJ8#X ]ݏJ&㫢V$vGL,j[)cDS>yF%n_@m] ZDw튪/'k7:5ƘLjӏ9flJhj$]Wg%oqNv,+?^oJxW$!)zwXO/Olb9w7uleKAPN|{(HwI̼ AEkDK_Am_\??o-57-I;!5^{S=Әu nU\gY@2M5+lC{ ){rs&Iɜf#^KE2C-d7]Pkm)jDq GP :ee; Ig%T^ķd;Ut'>*}TQW*0U2Y,:fZ р4^GK Ձ3nWm:n뇚q[?"֜͘D hf&I!lj@SV&5I9R%iR<ӑڔ-d<o]Fi)DT ?zi]`r/z:=h--eVZFZI (.|mi D}iء+j3pKֶ0VR }fP[“ӤJ@ej`d^fݵ/*Fd')5s\}6A,܌ X4<7COt+(@{i\-#NwCE(J|Шrݟ|~[ ݇`dES=]o[­ff޲!l.z}$2l 7{ 5 dHV0K)H% 7ZH!n sIT_5Qn˷k  P }p;j \0CBk %YJРB&&&Y H)t#NP.!d|,[(>d-`_0+?bP6LUg7V:6H 1xC#(ZF% :spY5C;Y{)reGkJzЏ)J+%\#d"$hC| 2Ɋp֝Y1bʖݙX]cDj|zaTBCNLY$#Ѳ9].eN-~L1㛩k"zRtߖdo\ V Z*ʈ)zS1ϲ''A )MLnp()xSH$^Iͦ1r/-.Y_R7)?"1px \+?@yO!*ukvǪZל=87r SN bz,Wz0mg[x"љ@R sGY*/EEOn&.v\ϬŽ+[J*t`~ϯW x`z_ m5r:?MCyE[4! ꏆ=<ڊa`aAp0F8}Rt&ޏCn*I03.L0!75[ULS$GE.\?NpX '&}>>||~Ǘ//'ɝO;4ן^>h$=-RtšO2W-Lt"X L|?*'Ե,{ mHPo"EsF֐.D־qHGd>OMRvSA*~P Ive {m(_HhWA!֡:!24O 5CJP oUV2_`Dr&Rq2*{Y;|>@)AAyI1Y oh9נhC3Fz̞S?]yq Oӵ;/,>hOEWVPL0Fa&=n&K $A^VUfX+rMn`@&Z]s3NQBklp< -- IrMm@6U,3!ꗈĥIeiSa74*f:{#H*jCi-QEi IIoFRQ=i\,EEZF^J_&gX-Qf$ `e 嗿zI+)+J#Z Usfg ьY8 w׾Sqnj U*#pٯf <_XZJMxgi@~+Jǹ4y߉WR'&$lHg)܍ּ{bt׼߄*p6:|p K: BB R~Sj^ zE'wD?7}wԞI3Bީr'p6Ii,fjM۶o']ե6.֫egH˾SDQͣ"G]6<\* dr4lϸ`1ZS6މjJml,Ǝ,/5ڪ&JWQʌmg*Ɔmʍc ˂$*1/rtZS@ҚKfuV^qcևZ!%i =׮vZ U]ژ$X%)jb MÓ&zwO [&Ѡ V{^GFw<؃v#ow,hf)u|A}}y BVq'ChZHŒ;kJ{Bld*Zң']Xe9338cX0%cP26A`@- m8hTwb\(1VۑK(z*o Ijmw|.$hj>-b'6j2ۅiS -ۻvP[5fweg4r-ƀfP (\;sU@. _1/XaIAB62*%;dsAP-o>ft :=`vVRZ@HhlUOo#q@vz`H^2b rF-5" xw'op e=1FxBTE`!gO[-NG%e^9rD@X&PZ,gF.C+& YmX.: &);Qs=p$JRY1HZg=vD8oMi'kcOϺV` 4gmXPfp{qm&{"vK[D;#RM۱EZN H\c3`ڒ[;7zV]P0YX i=+'{X x%s9 ;,Aqlm$KZ ,͖ >ʯ'>ש|%J2Zg4g&pW:-< }8U( kDrf~ٙw}y5x' a?:w;j'0DEiPx!5L:63С$ӑs䵨Oi#hRmM .JE]Uru47%:cigs=/J5Nͱ#J"D7l\dG$кf|1:,I~i.mBQdW Z'3H#0ڪMh mgSV#㱖RXj`z5‰v6Kc"˸ec\#*mF +z;!u0#  Ay2JD"Y:gWQXof^ KPܮ P&7 VF! ‹Ka~C(Xυ1P\ײ* /o{@.Wmuܦ$0*%&Fyu$ <)J0C:DTYbX3ɫXA%Wy%8n@LmR"g:zuC,фaE-K(!I=c\%UV* qqe4g&~ޖm(ڷ9,rj@]07'#6NhϴH(g+7Nd?fI9 BQ:*:V=ZK4 B7Na~)5A RON;mR ShZG2bZ5 (mPHbNQ-̒@,$xY9Kd҂2f,²h랏@*^~-(&X#[T7{Nlemmφ>Cf.%0&\I%Ɔ)F\I-`1ٶO*/24nq'54- QzF(u}3m*3sX7) 4IվJp}[%O5e /04$\O UE}[hXFkA[U,I[} y 9J9}MMfJS1 RZ&jʰ JRT=NcU|%$Wy%Hn76ƶڨo=\[3[0)ZK 007:,WFFWz'#z'D4zJVu/D$V|b'ѺU(8֮_.Ruw,(Vc]F4fT< -9@6glf,_42\&-3tJ1Zr#׽Yki7Bk-~,tEd$oZx,XEjuKUcĩm -(67n50HPbu@(\7 ȏ/~/$p nٳy1{K G*aئTh9t5}mG3s 3ouQc1G:.4r:D=U$;Ric 3!) 2|Y}vN1+R4h`-s~&`ħe;3^eE&A5bJK뉲p^KggQ(N(2Ў v׍fXQcjכjK8Υ 3@) gT+V` K1 NiJmJVʘLSRe9A +Ij-l2bM*1J4a45>M8t7ba69)M0ꀢ1a6ecepO\(aMB"@"Do?;Ē kLv3T.`00v\1Ib,TPei>B!jo;h(w&ԹVq:9= L_&|ځӌb /1ieOg J?NHD)JÖV틤2s`6(e%i1(Ky9*E_3fdBt҅i:PjN~>Is8hՂ0cvS_C??]WsIJT"6mTn3 dǑzYSR1fY} Ҝ-JPS XExOa`[v҂5myue*o:P#*o:T.n#Ï# mmmhK^G|yJN-KKBаI&ew\wwiJn›tFt<@RTg'%22ɀ RLNkh~),HeC}n>5KSp͈Ge0ͭf"+`{e`[Q")``6x=gi$xr}+ISoi|`}b[ hY` ) W,7yWoƓw@[{Iw`0a0ʼn_X+jy,>Jףn>9&?>0s?mD/<*ܱ{tnI,06vOA/ #xV? i2.Cd|aYF1cu/Q/yw(=^yb(y1oobfQgic5- ,d8\>EwyToBmO-5Y{|)t|Ypi~]2` ǧHO}`@?:qܜ ~(]>sǿ^{ /z0_u݌_G_aOWx<“g f\_3[<(-8zKô&l:uY>ΞM@֕}=ʁdҽz.`u(~O3 H+!#8{>Hk8; I~~6c MO(-Ţ_W2l2GEdG)Պل&Ljw ~ OzEQmEmcA әCX/j4۹EÙ"G|{$a%[D93sNB mPYwxtl )TeH̢бXP;s0={|qm2u޷Sٚ vzw@~eKL{pwe]^¢,MFxYCbxlB3%Li2Z*fJָ Mi܄Mh܄Mh܄ ݕ~$RHy"QGH K`PžrB 3w4Ōݐv"I3u^ gLƀ)wܥ賳-:XTwCk[+V}1AcZN+ 䏥Ewg qğ)3JGHB\Pd^&1 {;Vb1 ULX0(^C5k|}1lWK$YH7~ҍtgtET($$I\LQHxg!JUP-xy.Jt7ְm$: 6L:y0^gr|V8laR(bMtN=TZN5ѩ&:DS8+SeJl,:U *B yKtj4:U}H%MYIA8.\!A%,x` 6X $[ Ha눹"q$LOaB?յb2̨ER#$iݿSzu_fR+NQo% }T諹vB_ye/Sar$AfN9$ MM߷jnhj)yNrID!DrID!DT%Q寙}i)Vn'F 5wPs' 5wPs'*)CߺUjЗ1VDw|As-4}Vws&(cV;{O;UEy7@#q.=(g74NqZٞ <bYb(=IL#eQʋ8I" [T6"bKBhjлBZdeIJnY{E1M:bNm'\ywrKMpk՘[8%U"ԙ\ P οMQށd'>TF9L3y`*383T3$a` F=/ [!p.#pvW1(G`MO0 /E_ '[o(8@$`^h4P_|PrpVФ})ߎ_\<'8P3 8P3*,C+kQXp\Cz+֜~exᠱiM!d޻ܘu< -[D2$. ?hyoVh^4<{# kgH!z;bWww!@]Y.*wqgB"狃R,YJzeΐnU}KWL3ۭ_:}!` 4h" ժjSys-';/ ;RݐT32$-0C^;`˫A{|u:/([6YׯY=M$at#cȡfF J@MݢD2ISXU%pa) %deh_iȪ@q Lp!Xލ@wy $8]i̔>@70p1fKFL6O1]J6maIK7^V TZZa`wOlP}cDl;EY1esԌ MM~--d^{P}S gyxJˆ#T 'Kdk7ߪdվ=}ߞ-\tߺ$M-$47p!j$PPfÉ>hɓ&4='1xLs(rIa^&! ia@81#x\nd+ TCX JOTs~n*htK6A'F6|S hZ/09?=UI=;FQ6^([R.2΍x|- wLf>pp $76aJp^w2jOg:zP}3L9qv}~՛n ::̟>}ݳ??ztILbcsCC^ԏcqguTׯ{kg^\uZ "uт[ނnO^ Fl= "I nopg vM匋y7A6&ͤONuy}}&a  ;+kuCr9?[ߩW'N9O݁co ^ƽ/~򄩯bľ j):Ԟ:/!z>ۗS -_gTT:ר珽_މ ϥo5U@ٷΧ/ޞ -i/߾{ f~B~n|/dC^]NenT*G }.ѽ RJk.O"[WI5I%~}~/<;?Ng"{oylp0qSssχCОΟtpȹw.u_Wߎ.Czj|4rpp<fdY i L~u@wlM? oEz;M<LmïbO`/ྤSغ{' uO?Fʿ;Cj/|:I ~ḙb%pfסtT.ƽ'o(LKCAZXيg PvZ@ Ƀː\ fH0Of5X$Q~LvښpgҊ`/-^A;2X`M1!? rA&$؝i }?'gR)g̤~>¢MppUVcY+l3a$}7Me~F_2Og7gڛ_mMg|pL6FB㯶B#BH.5C[. kFqE]Uʰ]ͮuE"YhЧ2)oK_߿|~pVhꕗ[;?k7x;߳Zbk9싢8r.-s9k$ ѦbAYce G=mhf_-v_-WK﫥^N#ʞk g,HB./+c zEIjKč|[*=7y[O0O W'0PYH.[*DMu#%RVQr$qu41'[5oɛmi_tp!g;7CHRާG+ ,&i88-NF Hl,P\)GWX:I[AS k$[mg笖lG)-|k6뛼X4:W|Esqr6x>+ŔARdz6O۠}|W5?)7l;EJT\7~=[[q}cŭUBWYZ: p9O` t>c ЦHmcm!: S||I\gLaˇĦlf :6kc ݖ.T֞+q<2k!=;Ic[_,q_,5-+VҸOZ[Sꞙ_~H~*I;UnTQ•"%~ gO..N+h_/?+dWEQJL891jd'ӯ$a#LV-9cCN< @f[ |=7RbXܽynu)OR%r29@IuLpS} h{u JDLF҅PUC9hH`־U/d瀆_6h dF6M;pe]Pf lPvNbymut*06 4Bai}^ڜ%R*VASfFaJS^-{$vkw&JR9eEZ" h cN\֗ /iy'ӨKZ=/[ eX@IKZ|A/[ \kmr8y}*ؾ%[ v>WCVc$֑8g;T>~I ºe:fIvy\C t $ &iwrIA }%{wSW@C%,+$]4〱FpTnL ~I$THlUM!Dq~ZN?ɀS[ O)X骞!zڑ\7p$GIG0,PS|YܠDek -pWݽ>ZCe@jdg'?p- 9 -vysBf9G+F]$dw%wSEf.J`".OdR,3J%>HU3﫪 HqCf^B 2[G۟J%9J*ȸ>[cjʜ= .2ӋhԖ̈Yy !Z{=ʄ^F6CT& kgx+N7xnL+Tm Pj;j1Ǔkcmw?#[XOO):q%z霅CLg j/Z&PtrjY&F-%P?2QY 7+LB꧐ARҖm>&@"򻡲P^ ie>Mej4֩L$tCKo4*s= *0PBpC㙍D4ʮ;4hۋG"rDKxAH6zz$P+jJ7i-_ $aL$4jzkTH~T$9U 9Pjq`Q> M WqsCځ-߅g}O7'۫G FZ3G9::z% mSl<"ZW)DtҴVDkppZJSU \˜QCʌB%P[ƀ 2=F^;٤kGIa0xJˆiQL+$3M4 4F7\\Td~Tࠢ otXfG1̚ y$Ors{?:m/aD]"r\Hv WsjSѰ$LTbV >B7q+YHfk- S룖{.UdBj]f蔕`M%]3HJDS> :T8  +3) s4HvLc]-o5-[~o-`R:^bmRv۹LE3^bꬖ@.=H_o ǘ{䕼Ul]ђt\mi}gg?Iվ5ZwNibt է9Wݭ |;}RQ_߈ڧ,u[2u^q U,K2p0i)s]鑆ke{cXvi$CvwǗ} !*FrŎR.|qQ2LJ`Q!3WԺLS Ug@]U)Idq~j;kh{Ti+~.B (2R6F^?ZdkWmִ͝*9xk[Ǹ1*>:R(nG5DdST5+Dl`(dgw~g>eA)m[)p%3vЁ9d^247t&XDwk8G&Zң 80qڽMA-ɜ@5hщ5mLAE0K͛i7y::knIyI0!e)E(YsT<ZÕ[WMfw\]dh u KsgmH!{h~?8༱`oAyHقU͐ԈJCΐ!yuw=]UGE~poEyO*{y3_z[&kK:go!lӾ$^ui\ ّcX (kfHвP2J3f1ɹWCЩɇKgRE6LfJ`]3݌i4i-ed7p^dp29YR"Kp/Spbڂ/T@88|י\ b'`1: ތ%$RF0R.ivl]Dov;Ѱ bj,sa :,(G4e %p1i ڃH3!D .OM8YlȪF@rsD|ge7X7_QY k+ ZTu0*ME:RQm*`4+$G+VqurF@%,j%Trq5l C|Tу;BQ[9AW#,*2gVpAPLr 5Tk-!wPx@dZ%MFpeXnQU"lJY SbY>rzSCXMb=0Fʭ郈V5 "ހ^reM lg1A,M酥]B>ZDD< $s*JejLS ,ɌS Yу75Aute$Z_vFu#U2w?24Mnd,$bI+DZҡ0P0R\KEsW$},hd'kq=YbnлԨhb vH5s)K;yDCE$Y摩k&%kH\O9%-ի;b z2ΒuF|?Vj0,L>KfG!x+HrKA1}<^F$v;#s+T&O>Jb ZKKEq}>㇞?W$' фC|8,9B"}y81T2Yx)O dWrQ$̔VHR6yy)TRTP(9ЏHHڃF)^~xyj3tƕ9t>a|Ud73˝k@E&(&|ȓEt#$f0C_rf U|2⨟i gd wy()˵W|h (JhD(odLB۳t{ WN/yL,LwGW?}qP+ PRp),kiZ[hON6n)alKYf6ɦ}ijEir'52qRJIR'F#X$\+Rvhd4l\hmKh"7:-\O|q( ձ[`&Pwi|JYBrุSqc~&h6=7ufiy?6WkӜ[2n7 fY$!05V:?y9 9i7 u t~ckXKi:c44jF4U)NRÈaLhN3 W SpjKF$K _Bbѩ)F16upؖAْ뛞MK[djG`?F߻,QsZ s njǠqmyw8*p}AbEU<,~?(^h~ˋw5lrmH~{甬]cmZunW[;'"[f4 {5,ZSMbh8 1t;YQkíߛ!C؝ULHj^9|SLg]=h]vڬq@$T y>|pp6Yr/,?5p{seͣrX#J+vW~^ze(PfY缈RfQ=Ǥ &udd}w,eYĬJYἊ$>hN;3k@Wþ3BǕ̈́ 2Ίȹ' &CI1i$V+¨kjٷ*8F _dzpQ/4R—ͨyK"i%KFSyèeGs<6 V48nw^Ŝ#- ˅EyM9TJ/K̪"dXF"PVw4uh6CRpKb55LiLّاZ ZgSQ (0rVSJ35y SNz O6G `|Ɵ>D'"]c|8FDaknK^ߎ,:aPMeL-F?? ]+3~vP}PxG&D- 7_ 6I)m5m:2Gj`F`dZ+ښY&H3aÅX &:_cyiM",cINY(3fE=ǻ0Yk-; lQkQfN n܂M \hLl|n4%Co0r ۞DOƗ4rߎToLQMȎ$x}]F3O B30)@kef'o`Ubɼ*b,puھ $Žw[zĕE}|dYxg/mF> Ɣn'd`I .\sj$0 ;.<'+.Mx˘l\ %lkն$WhzZ'-qY0%VX?.ґJ=꼙ą)_v~atvǧ}IǓ;MV?Iz6Kq,9Yr\%?k%F&A/=u['H"pV(YLC D8g$Ku;?tv|Uu(N.@< $3R~h<( (TkgclOkkU;YJ- GP&f&Zybdb-#ci$5qZSm {26(Z2m䩚IfTH~'@8YJ&_ۂtrɹr"Pl aƜA]DWͨv+gculA3BV}A Vm}B2![x ݲ5 íM06e 1Ɋ>f~},;#5Lrm;Z |f4-Oو6Bߊ^LvD$WVryO'9 ?ygΩFq ѹ/+L^b9ӁΆ8p_N"] R?Y?ă=R 鬀iQ6/F#x'})O[XdLc(ïPa)J]0酏rv[x%nL?8Hݸwbt2ԥ$T8 ?$ 83g?p</?9O1Cc x<I,^0p~rO q!@,<>|ÿã!F>m>f_'pc0Lf`pK5<'|Կ6l{9698}F|uC@ќ6x%܅F_9^?Ň@,42ѓiۓuOBO=iK=< =e~ JjMwJo ?UZ S-O3sAt/|6+CRrOux۳X/l Z{Igġ[Dc*C},yc8^ᰔn|}?z?_1`2~?OnƏ/O] i9߸~ +G|ty];G_! Nt'?10 po;uuWx_yGZSyla.4`1I.)b>nb!5wBoneEȯGgAӿ5L&>˃3{\Ӿ9eC@>,tI/ݹKe?_G{o/޾=XAjR;ջs͸bb+#$QJ d'#о+^;-cNDZh'E(tǸ<&"ٻƍ$Usw+E ,.w  ~q=rAULQMRs*"g b_^^!霳X!g,W;:hӵsA\W/U"jcLWvD*_Ou:M}4 -(bybk:9b8܂ze q<g"ո* Η?𸨷_Z<_5p֕Ze*,zwݞarxxZ%bpq~!_(u0.9vp+Gv6ţ>OqiVN}1ӻ^ k-8*6wgK6^bͳhJJy̔*KTPL2ߨ5߄ǧE 95< 8XgŞInazMop`V.wIS)-gFѠa*Wys {MAlyM8ѱbBD6(T b&Hkd1'zI4$wFC?Dc:lc"Pnrc#b['](.&Rh%K%.GR`Lg`N8u *T.`Ai]jbZ16Gg B0]VbQ»DjK]O.AQ3׾rJJVc8)@ K=^ d0F)drfƹ  }1lYxƐޗ$ {P0ZRX:0!lW^pH , 2ưQ/U\#%cmj Q̶]c߻XI0+Bo1T1!$, v*O?.?MҌQ8ֵ۴`H*Q-%[of|D3ʅr-r|7oXDhNF# N7g x=jdᗇ4VD]pj%kMa^ e4n?n~38?Tu9wuF~gyނ£T_O+_XX R-0/H{uuiOSu=H{E@JD5Ng' GYܗL҂{6y y× 2B8\cԡ:]][e7:$'<4v<: W-xZ>z_cHE ,v`A偶4rȄAJ&ռQOX'ejEot 3?ez3)/Rf&>*a1z? 2q3S-/}x,?qq4x]qwJ}|,S]cu8м)n~*1f@;^2(snS5^b":}0$)MOlKnCH\D[2BbDGo{\nr4qTb^C:C, s% a\!)pUD2QHUxAxhb6-$$<+SI2, N+ GpOi+B+n,>V4z +_-HV>{Eٚu"dJP1a0Fj+#妔A;C\#YDgcf ")jvoПټy(=fTؚr[8RB( qLԪ.}qW̆wMy髀z y uB _0؁Ep/`!KqYY .ExLZZ8j<#HFӯ+ Ғ;4BaǀYL)a1!Xނ). *:R*83B|qz640Z/`|Z 0ɚ꙳)\ݘ}ȫptW{x~ ny|1VM)CYlĽ=Ÿ'z D\ꛙ8@!(~ ~ n<3𑌄c4HƟ !%piz F7!O%VI@;=wpZ˝ee\҂LFToZ;-@3l4xƫ`k4BRץloD҈Kģn<B7|M'0r% N$ϸWF$x}'E&SÁLS"02TF  '0„(,Cq$V /'(5  ' GԙB2B j&c Xٵ mro8rDAjЅC@h`vli$lAL>\]]MSؤS4Wd rY%c͐/:Yrf5cT}gp%DZ: 7R0ϜESĹt8>~XZt.E",·妱 wbjS& >S )r.wr\xLH;<FPll Bra8xLPP!V8bS ȮQӌuf;&fe)uV6bFK4J CSp|MBeZy&#£fnCPl;*}H~بJ{ B*E&9TT:h8=t&ǣ `VE#<3~TEEa !5e 3RǢx-DQ*r+*}Q'IN> oD7(Xptpc 3&5^")Qi'xAݵXW-J@,fi7LEAv#>$G2E4IBu_JXHKULL92SU0,x_1Xu@+i8JgC|# g[h"t&5L6.]>z͏2*=M;H /*tyuNSuT:tIQ)Q6C׺wJ9Y2yp*s4eQޣ3J2CQGJTcciCd.E"1*axL_Q4eJ#^J'a[T:J'*])y-8Q^750@*(p;K'&(?Vr/Nb)$88S$8ޠ:Q4 .N:L<ii cGx* 㖟qWuyCƲyF=W-/TLЃUîL* ZV*Qku .A9+c+/$ Ya& p%0sEwiaDb Ųp)}'?{d 򀉝A&t)8XFc؈OWʳqn~ec:z⟪d'T'%(~> X}I./?x RIY<_}ws`#"N?Ɯ}q_?Y~SZ&0'w3 LJg[K^/#7f:aǰ:H#+lL@A`EqhJhP.zS|kI$9k4asnS8/}()-iHtXyoww0UJ<|y\?eʼ=?1c~}L]/'uLX?Wo>NcccZ(lMg2۵eO e 4syC.pwh⴦`90bcOYǽ]c҂]YEάKY$]_*_k+r$VQ# \bS$ENwIJ.S]=N p8 J_k_gT7qgȌb])F6I0=7rkQE箰Qɷ6: oʎjzJ7S"ޱTswu&i5X@ٞU&ӥS$UejIi@ˮ*ٻm$WTwK@7%Vn6ʗIHq3Zrv6(ɦEJ2L,dA tc>JqһVbf6_4TIϱh~=}?kH ϭ /bV}bxgZV6Ud(9:>tz)?)'e=!C4 SFNǺFox7: >w1"@yd@B8D0G :h|@f`-$л77FBsﰽ};Cٷ ND ~VjeܶKi U]:%_|RnMĦYK2YVw1FӲ`C0|LL0@c!Z3N#F1 4.IjAu0x `S` [ 0x\-5Ĉ%5 ϟҙPX@yf*1D`%ExƋdˉV˶ y\hsnwQ' ^=>=>=>W6/m[iSt%yΌ6%q.,07 JoO{08 ڄҬ!Vw"I_7@hc} H!%LV ^!EA1M%ˀJcR^bJA\~m4& P0E4̈c`?ݱVE ZC*k&aqY`tqp U%F/5T,HjҦD| +NK;%*u+NˡAyѣ ypRHF&/_W6.UX \ކ&YIR)RbB& #r/ÎnT n@AyZe {>M[tmejonλ l^uf1(:">{v!Pua\㡼{zPtkoz.I^,\02N:6lʐGb $ʱ_E`"SOCֹҜ1O&.9T$9:K]𐥂J(*v'0T9n5NBtiR6!ip#(ղ,KT)2Keae!͕biZѽ+ guqw2ݤ>R.zP+e;KGՆ`}B>M_fi(}} C4 Sȏ~%nZEX@'mxсλgz>!)%>n֮~D>w-B[|kw!o1ّU$ pAoe?Q v5S ؉J-\i~Fpy N9s6ܤ'lkrMzOSBA#ge!:I`sVF\Z d*+r!KHA7,o&t-}'^WsP7nb>jrg#~Ɩ7Qmjٹٽ_kWx9/Oo\;r}PLw˯BO/Ij&djm\s<|[-]dw(j]4{5UT]}cw+l9ت}M('Kj'VYL2P6Q, 5 r#p띗=.Y-uԜJ6g hZq|0[)syGx.RPqd)bB9 }T=ܕ0}@>4IԢbks,O Ӡ JU}>$rjd7>{E5P 729/$k'p[\B"߹wK)02 9\4}w?]4w+_TO({QM㾭C:=we7O2n)'g6[ |ؓfmGs_ |iZNT*gY%S9۪^;X{=)1D/%%/8QSJklʘJ9XbM!1'cUm?J9`,EM#t:Zٝk0J2Y͞Q*! TsFbEEz&_f|4˗DTcco}M]MI(&>-&{<FXjy *,F'.\<('\mm6co) }aL~A5X$}l!Z(`bMhOEq*ˁ%Yi)&NERBr x|Z7u)vA@~|X:9an}M7۸\!Fp:dž(|W˧Xvb"&>'{-z\.7]K6^e7aLg.Fz1Ɋټ(滿I+|^{|^{|^{|^񹼃}Kr'OQ'MaHeΤMWYJZmDcsm0G?++\;X's_³{V#eaVЬ%5ރqTaqYAtqh븅keQ\dtrʑS0װR]bMubsH2k'ѣBƨYNÇN%%族-V ꯂ*P`5Q%(5" K%FYRJ[>'d a{LrJj#Q䯫v&"./USjkg$J!l yoUd~E1](tv l:ef͢F۠,AjY/-a I84݇s%=W"(ǫi/ Ydr9sIj"1hA(2Yx*gk ?!ٚ}`l53Vj ma=iH sg =]qJ@?g2vگe+B~-pT*<< 5[+(0UE@jD<rVbr $?ԣ`9ShY)c +4 K icJ~< 3݋PwC^Vx2[ RK=J NQ!|w2e8S9Ȝ,gJrn߬( DТ Ԣj䟁2*,WeV#RrK$Tf2gZϕ y)V*0he6QUH|sQ\<òpC;#D.89hO* ȲB@YemxVqjjwYՃkQ䱼z?.ʎzuT|;')HR˧7/_|CQ}ϛ.e ~Gϯ/Gy͗c}xׯ|Ͽ=f|Zm:R{Bkظ:#*br??v~?e}ty9nDuf]MvSW ! ҂ݬC./gL!t!X=᧪}zv6%A }ey&QjCjYU{9Ŝ(%bxho)u x]DCq$i(A ˻ j`]ZF6 mr/WF|#TSbǧZ&a8vv[e?jO(] {Kk+TwjqV.F/]YWґͿʰӖbSgWWA6]*GVaA"eB X2`-s!Z+ { }T!}NW»BԆ]=.5µ::%Bjps\B\)خ. -_ҏ]mHmLŀU "QD#r+~$Zt,X!+.jz7۪j^̬=f=9yS?3y:>}sKUhh6?j[yn!,*vS=2^_廊 ݇b^3ℷ'qFa+cՆԊn:N3Zyb/n)ޭqFaJލډ-V!Љ}F6;yb@B8D 싈K{&uGrDR^VM!Gң l YIMZ[rң $TKym#[T0V~iJ VfR,/e^˔l6 65rԱ|1hC4rG3SV)"1&!pQ{kn86QhW J Mma/:>cĺeLO쿦7ܻ%M/}|Ml:Q8RB ^xMӳ4|-?$ZKQfmWYVwwL` bUP>x/"iK[L҇&1 %(Pa.~I\mmrhҀ;ۨ[!m6H4#\l~l,kk-\&>'{-,mK6^maLg.Fz1Ɋټ(滿I+|^{|^{|^{|^񹼃}KP²*Dʓf)?{WH/"F/s!br͗g=c-gv6vK[d_-y5X-X,OU1o 1j]r²%c*T8j-W}&Vg}lu"8gwŧyJ|{|> Ӡa:J?:B4gY+uhȔ3$8帎 I R TR` Yѡw$%9 63L;z QFq{({3X"so"2 جD ²X}h]䜥x݅J˛իK2I|lQ+}TmteJ_&$QT<+|MAXyVZSM NVzV*3}} TfҚjҌNVzVJ5%qȨ|]?ܘ LŸv}wT'4\ h8'l`ݙI1}K[q &?aK[•0%$V+; b^^B90VBQ( LKs@ʊPxǫ4ȰuvABq-HI/'u6礗n:FngBxD /CS߳CLOZI_:sC| /W}j]m,Vu&hŤ ubɽ btQ{+!(λ>j'/#ܖƙ(H z3X9>/S՚.[%JaInɚvfM UXMiN:k琁pڇ4oFH`4"4v%po[Nf Q9 ;|V$cp bؼ}]Xc{Pw!fg-G2\*C(kpnВA2Ziӏu3BxvVvBli5Ŝ^NY2x!lނഇ| ]L@*:CH\QcD[$.4D(_y !eMN5rDOcfp i a/Q~j%;YQ[i^2郰Ҽ դ -vVj2̓FIvJ w)L!hVR1MqDUwPTLQjA !v0mkme ! Cb!R(GNz-9S )Ofa,%vݰ(Ɓ;"+L+ԾߤVF5-/ \6w;NAuH7q)`߃ zS[:cKD$GCtqkެoyX7woz*@d~?8u+ŁRzpփ1-Amo[H; erG- bO^_ᨸ挫`wϯ-c,T&FQH\aB@^' ƃ_Uv[B2 ImFTj󕧶&Os hN Re59ISS8I֋ J.hËx 4T~6@1W[7>崼 #|O::eCIXVԫ`]N`S 5dɑ;yC߁s э컧i5ꤛւ};BM%$!F s1`_2A&0gRH&2>0aLR&+CO>TCO>4I=>1Ā:8l$J[cA䢊B)Hݞz'&c8TT_G;9ਾ}j#[!Z g=2ȑJ-q>$+# 1TQ֤OJ39L99!I!}j'^Ծ,짬GD,#B L',2ݭE4]O-Msi%ƶfFtȢG(2J&?9%p;f)KS xڈ36:gw; }f9糗&hԫ}y>"&5b7P)7}>z半{쬁>jU"/z\վ_y+5/ZkN2]bTޢE[ahFl7}GnK{JطF;j\5_ږ#ՙE)!NjXZJBfJBΡDrdV#Am`5zsי9Bry5vݎ波KJCN9QB/iX^Py*kpvIÚ|j䯇/i8diA`ppJJ<|;vBl$c9]/4;4>^ICYrgpiF+i`IpU0 6pICn*`EMYHne E{E륒9OV01U7ORF2fLtuudFbN {.q.#J@V*sEu5n T$I7>l2YmD6^K6]mvj(GK[.l,)=aɞ_R?.md QI7ªGf_xWsMWlܗ伮Kio$7AIGtd9Ni۪ d\0!M>.FHmYM$ICI4%!Ph2, "ɵ7w(EbֻNVrs4)&)KgI3Ti 11mT<3!(]-HH;ڊ :*SZx-M;BsÓ_(3Y; 1GL步"rG^]P'5>Ys>K2(giC}F,m>+Š3xJcPaNoiq˩1iAA}TnⳭE~P]L9''E8`E-^%tl' [><xfZNL AK!'`v)J<ӊɋGm,˒o(Km/?$0^o&үUiY(Oqw{Y\]~шgUw̫5/WWU 9Wj]_Uudv6YzP%$8P~ eA|CR_;Z 9ojm%jO@m ^n5gGQ.}8EA>("x^bi5>3` @簸b?٧l6B/Q~RRkyJؤ쉕j8 .d]933tXIؘ-Ы돕^ r|Vw H|NT!=Dk ^s1􇋖rk\<~+ac'7Y*Wa)rp7i<\#̞t'*}ռiii -FxvZ.rgàHO^u+bC(;y| (`$;]P6㖢`Sou&zZhГniȱs`2HvQbuD-$,J&5V2Z,F$-Ew_ AC ,v[#@ R^FIE5c#B^bR9j /6Zc@dH2yUs Ȱs*Ti@ pHxfx|ts%~&)0siSތeD/چxf&yȕK2\Z27{ )-%Et@ͣ!C6͇*kaqkU57/p5zx'B?<ϱ ?~ִs&?܆\~ 1ok`\L,/8Kn5ewҪ)挥۳Isɕ%ͮ+y͍a\i%8*!f+AxKzVض%-UWiQ|͑IieW md4n¢yށ{#uo-bGkW ߗ( y"W ?=kzy|>^-ne//?-}l+lUicvv,-vg5 ZNJL+jhAN ?Z&I*13EyZpIڟ c$ {ok|C[AAD*~{3gC2+~w|AO]*G1;<Ѧ=ѻv&w?X%t-S^R#ǯ'q=Kn%Sǽ&rhCRݔ1PmE`h6j7sGn F7M 8GBuXpom8t5j)~k4G;-H[2ao . Umw81j bI,Ӄ$c9\"΄>32_ 1y 5Պ5.RSGYjYbYY|uxpno/no\b^/XΥ_ָL}e[Us?}kՕ@xu^=_W)[LgqX,uu܆mUyV:Oۋ{tM1,\ޔ d!qeSIVM7N'n}ec:}ƻH 꾽[,`wB~&˦m x7 Jw+6mDNiحFwB~&٦43u78dCKPK A:<:$VTrČ/QI1 ef,At 3j,Vfwmql%8nT0OIeǕT`FbDOA! f֯rmыVkSSjTa U0%VRKG: REnDŜ=)O-Bψbсw<vI©#uzM*o8[8$wZ|Nj$Zĝ^^d~/r0M [NN#@Ô'$zj'zM/HrSDQ:]iUcu deP \yoZT7sϚ>Az+h~!懶h`gVRKY.%?!iE]]9Z7X:ZVӒoog!mlqDl# iToG\˷EJ3+hJΩQ:[16șT8[JUV5Ԋ4i]% *QM+Wqu#uhL)`@P?AɌ?.yt`*74ccTXT+)z*j9 :XϘjw1w`c7ֶaկѴ1ZTRu bBs2BZm#j/S|1*MPrgzKIҧʎCyS},[N'r[+3e##ŐLmP֪df\XmT;9u>8 OMr-eRmOޜl0 3lZ>oW&9s?Y_f֔viC_^^+Fn`jo|z/IR4ԅ'Jx=VIS{ ;n+CUSVA^YãʣN<✵WʛNfRRhۑond`*CѠ8i'0U]PARqo9#+% ERDw5UqMD1TΠ*J[w2*Z8]eVY)uǵH Iu$# ƒF{NA8Ze oN zhDR%7SݡP*'RKEH8)]R-Y+*OQJ%ܻZG!Q}CfӖRqR*dZ0{)oWklb1|mZ^+V^́eSPf`M %BBu{n+(Ve5[Wj/ Q+U(e]RAM{`!Zy1/[0J6ފQ[cN5 TS9CTE^馂 %Ѽ! auvs!$} ?}l[{nVyG\/-a>+L2$SvOD iPrG<_ޝ0iAx6vL?kz͐nٹ~' oϾ3\rsn[kS4xÆ<Ko(`8]?+,l$AAby wz5|5{ T-KhѐYs,Q}C55IgZ@ K.&nlԤ_7PЅ^+r +&{Ym !'YbyZpN0Qb/L IUsS$'ҨslqmЀ.Apױž@1sCi":<%w* B n,kKPds>sѺsOΜ{+ԝjMۍξLwImA*&A ẘN}"6+6ۇp͒)$n55yB햋AwhEl;nHևp͒)8w? -}Gvҋ@Pr4햟.eJr '(=.6=7fz0G5?Fh1A2B5o]eHohoc4|hqFgqKxP|0շG~2zQMa|#!f3є2&`2-A (6b 9PÀ)GS(%+4qjLk@X|[([P{Md- C^2Зծ[멑%XR -_ )S0%g{آѰbJ'Lc :MiU(5[Vs-Q:{̩'&L CˢU uMɛ>dJ;ϐFh??YGz!Zw">~ X]6|,~VsZhs;F[s}.][Je'm8%xK~e x$ȗ,]ZIO|$OpH ) +9S]U]GEM| <#6pOeFSyz5l6Fs^ơE D5ۡL~4bޫLQh!lB37A *&`$kQ}2x2FEz47ֳRDPT.pd%,Rv%e⑥Ew ݠ͙cV"s+bɒHpx^R/` /ǡ'2`n6U{2t5G)G>L( G#)|tJ45e7; {l6O8%9 <Mn~vN=b:K^RR!<.} gPJ\z\MrLCtϒhhAIʐ,(KS"2x%܇%9p)roA 2!lejϖȴ*H/A}ن4g@T], %. ej̥ͥqz(<.֜34rǥ<YuVP ̥'ͥy f(@mN.ePC3F(y2+-,Э3UxuN7 /J PVb, z;^Z#3,u}& n0QљFhLA0uGg&ix1!4s7a`ߠ |7odNݑ^&!q'P2-ijpǑ2R1b3NҺNɧ1BI2o͔5D,QkJQ Ryj.fwMY`5v[]CH-yJVv9!l[K4(A sjE0 w<,fR%hް՚Th$~{ 麢@jrm;a|tiH%% 1c^œMRB4/fcw؊"dt3zy[wwo|BH.[ x]H8n (t(#rBIKBNƨ=#O.[Dࠅtw7ds9eujB:V(v`ݜsB}L%ڦ9:oB:'M )ݘJiݠ*B',әt.ÃwQ^!] 6;޳{tQ^!]ZT: .+6DҼB.I%"Qp)7y\ZAm g.=E. /%Q4wRAGFhќJAy &A#UK&E DRw*[9ϖygK?@ -JuKQ( ej'ΥTq)f9<.֊?NK9RΪyJK9 jOKAq)T7QTǥk:kR yպT%)N2օ+^QǯuU:oŘӼ؀=MpSVXv]T }Dy `FP /^w /@[ZMKQyyK 6ld弮G!p)jB3шK* δEP4*vBF *%DS#$Mb{T +{KxRkK4̪8Q@}نZSswW\1Mdtls|N5D"D("pMDRY{+z6J&A!ǃ>l&Lթ܊XRГUA>O0FdB !cx`BHQFe ApF%KMxB( *pS4zA+uEl=y1]e!Pht`hBohi(yQ>&TF 6 K*5=]B@iOCP HwPCJm>Tge(T34'Y2P<-#x[yǫK A:QÙFh6huv u19e Ha2X|x7'[]%,*F6ר~,\|XI?oC??$ ʀvRRh -ejV,HHTĿ)%LY|OKE otd&l9F/X$aGzO.nWwbP|Z.2$^heW^2@-WT jٔ" 5$[q[6~)5Y6(Wmo]niRm0C@;5LJպ)n l2o>,м7ꔵ N2cKڅ@3eRi)@>%xx⠴.Sը:c1=1"zI5 Ly2<{PVz]?Qw7]: lQ2?ެ>f"%WG-{}#D@W-:ޕ>j;F*ͨH UMlIZ_Etv`J%IP--ZPa 06m@,bn()JE,R =Jo= uJ}xxbm8X7oe3mFDVdr^&Kdf#\3/BƄCCw苸]3;eiSF}2O;~;F7jɭw_hg;YI'rAYɤʹK1z0r&CBAu;Lw 7p;dG*&p8nU I?FW&~R "7FCej]V5_mkۢnT_ 齿~Ij7Uh3%fw?~49e,7`) pΔW" =xupFl,jFh$1pKg Ht؉!ZIʆfL oF nɜ:p"$K9ӨXITjwR`]B)Fz(2aB#2uUBRʐ,lL2:`oKG}(IįRxY)>QYՖa(:|I/3+zЂ^ 5(>uoh$H8JHkXqz\ vͅ;sׅ9^>BVĘ 듬?O֢ $#X$#9l lR*niJYsا%#CȘ4i;,_R HQr* Ġ%&Bb,&]?]Z⬔'S{D.cpP-:o\>Mu*\Lt5E?(-9.z5?oP ٱwLz?Y#0 (>00':0I/g''6Z$Yf\<*?3¬ ;T8n09`v|} v^LBL*T^UhY-¬Pmgrn" |NVߟ2m  倝ԶxܕincW9K2g?٣;TSMGB&&n^wVOO`9{~tCЫsЛR_bJ3v :KOLqm4hǩ`xœ !IYÞ8&}teC5=6Ǎ|p@w]~=B F!5H^83\@J@< D8pڒ-+- |0|k8xGK| u8잢|)"=4h?_cH(ZK[cOPsJi5 ܴMOd7mQq:F[/{ո[Sj[<٫$!-MM%Ai`A 嘆   ibWvz`)ո:3w]}arjh ¼9Tu4Z΄$: (d"B],tJ9鬢ydT1#LA"~Xz3>`il=\$Rȝ{˧--ad\A:=jF,pGDt2=d=W%c`ފLB')Fϯ~a$$W_o$$Ks>Ds5v^OdQjwdHY 24:$XceuQYm<ـ0u@j~L^8, c&Bh|.#ZMY$!jLr҃ A_  &aǸ#LoD${?Cr ZXZ,xc& $WϵPb@CYy^hNG1|&G@)AXB m]: O,k( E﫡'ߍvy5.ne1 rIn934;KȽ̓0dǛG.E&av?%\_e|J<,gaJyIf5U.r=YWnlVT-Fy xDU%MbSЃyN [h'#-Ƙ`ݺʠtw;.1noz>,+7 b6r_MBeO-agFKAŮ(H A /ਊQ'zD" TEΒHFD ̜9qgqnB!1tP d0PeH"G%^؟KHk\F., SѴzG :Yʅ#^Ê;Drҙ.a-1&)%TFV~ ȅ/ؒ`YHSPɞ70dd:7vhC+ Iiڧ70%#`!N%5չSIaBϝJT" ek:*lLJ[kުՃ6JJ4WPդwkW(aቃ<{cob+Pu{M6mbH4*6P1h$P =P0rji*.]tC߸>/vgX- )u]I@_n'n>o T>BFFϗww鷋ʹOs~n?Ovo&f){4w1~LOk2]]&`ɾm"PgK?]Z=dk%O:]h6Kt}Y"^-m/sHA jgR%B}I2$Twvk5=;谡qrIZ[P:1&JjâiBy#>d` Y,:c+%oOPjy?USTa>gͮ6"&㻇njijKD/4fa[% ɨ䱖C'l 6LPKԲY骓RjTJ{Ui3!hϭ^ZN ǠZԨ3AԪuꐒCwG*Cs z./ۈM$_"y-2Swb%D &g S7"\þzjkoPbub(O%n+<Ҧ=T%V) Ҡ`H]/XaW/V0&sAI=ʛ_T)Оw\%ihy4ųu* ~eQv?ϳ`@aJgC&#{Of[TxM蘣&Ǎo˩EȥzL-PJ,â}q+P&F11LV/#àGt4LmfA QQs5$Df O,'p0&:"(ﭱ*\u"p*O  9P4iQIbJ:Ԍ"OorKPzLDdXwwJ 9$ ]Ar4 9w{ Rlu| -e$xrIx9/^ArL05N&ӖAɕq,ĻsP-PlE]ZƹZ'NxT+! -)ZfsYDҧ~Eg ˶ReVJ1jNiˬtNZ𳕾D+J9dB&itg&,e ϮۏsvgIGLKfo~ BXf <ʦ]QzQ/ޤސq=c "kd,൜i0V`7?&OQ~j-|՝X{7ƚ*N}r=;z`!R׶u:-b=J+sp`@nݡ՞g&׮!iDiJoԷWt4Ev)kR:éE'"hSբ!ʓ|E*נ.pj@8Rh*UC$`@j }d!`_Rܘ ROpހ<& :@/7ٺ(1J-$ (CdTցA ꤵy"xIM H S4I,AEn^IOh=u1OYe!&IjTe#+#dLuQ?h@)Z9px_ 5?] y/׉(dxp"+/hȫ|j;Vn֢<mUIQmqT# Ejٻ ~(4ʁ>\+tO./թ%N$?;kv@uy R =y5.Ն hK T7 &2>:1f6iֿ[2HfaHeįֿ< %^#8BH[]ǂ?`-n5>l-e1&FFYM8tzmcV>4uJz_ssRKHR%PXog'q{ r[ {#P*D$('r!A%=lLH@iD9s( i"ոgH(QY7<2)*3J*b&NJT`"_ lutf=%bF|%1 TdOhjrBnuHE"E؝# "H\eel:݌#K>H2 c'd 9=*ZuпK_dנ*f{m$=AEiTOd jRm냑z9E`02.ķhN wׇ?ҟQ2徠ABit.=Z?6Ŀy(~P?r[ 1LЩ^>HЬ"DY:08D}X r14dqGђh񘹈jAji"ZbfPFzLa>m.S+ G B%(yA5TH+t|_ y7K]}lҐ'1DQ+<`)I1łi_0̪/Pato$o?OF4r2huXv>mzigV D F1,4^(RNȐT"`%0/Ĩ-&#)c Q9 Y0N<7l)bG{ s*77cG$/% DӖ=搃RB-?̒s{#`-yKhI A{Z0'.=(hGO%e PrK$Pb>>DCZtc]Xr{]QƎ"]37(݀/>#PI4/bFD0TK)83]"!,,F|LȨ2N#dJ ·ȃ ۡQzw.Ar^2F*97XcfK bm/B㹌B=QU YgKWՎ[0Xj(|D&lG4ߔP+}̶Vhj^Rt @UvRRD4{ Sh#9. 3fDR2@_>PmpYmM':[P`7tw_sb11"J.bnVM,3S3?8C:Z){M}̓!+h6-L [n<6:S-ܲ*9Zucаò,- ?PF /%9rcB뀗LPUyf*^Ť懂= @fY6!na#o. >;hK5~fJE5fX"RFA>З,o,HI%ђ)7l'ʹׄX6l%[fyK lOg_zzrK/mZ°W#~߰y!Cw˅$D~ϫe^VqRA.d9Wv/gZoi~}Tٗ77(0Ia1(48CS U8 `5voLHDj+:2^>ѝ"Ph8[x4.9[>`L>* lo7Ȩs߰>sg9X6ePwJ8bLb0Kr4".PtF9)Z |9FhJl0D[B~n~i64?_^v cTW^ݱN)1Cd F,*orr(À=@aY߫(~@sƼ|r_5) ;ܽG^[q)pT_㐖z1%a>#20/QʐѨ}.ZViz" 8\QYDuU }K>tOW%P0y'qW<𶴙!K?;l:]aX![ 34{VzsGKB~C`fX25=`p 59@^ͪj(㼧\rr4eUXY(\C 1~ ωKγSgdqˑx'K"CaSL÷+"X yR6L{VΗ[mxqD6-<2c嚮q,CU:ӕB/3]g 1~x>LȴQg"wGf7427Iu} gTOq;PKDRr>97/cEo+E8m`< [S_[i~ɩKd9BsP×W`UmO+[s!b25hdӀd)cNS%uqCrp+D7kb\9I^=ot͛V(ZXgtc|h&qX+mm ױwCYQa5(CJ0{Hlk5R7y&77NV+;ق@B_٭ K7VDVz[o6/Wk|.3ݭX!5p2k0W-d9,.Xڨ>©B`$ra̓hbu4uovC_1' >3ܣk-x(R1X#4o({tU׼J׻ |*BvyõI<@uɩy}SkjAǩriߏi\ihulcձ5 >yҡ̝'=uK }@w"(AkD*dy[z+HM;JKiLdehaa64{nH~W{奬l^JE$I( w&`Fe2-kJ:8Y/~nޡu = IQLCdgrDK1Ysl yjt67q.=zզxRkZZ:0^FMBwm(Tu@hx2fдvxkV3xU>1&T . y1P4_ggkvCsS'oj7%v늃&m]LuvCsSVT=MPFv늃&mcFJ{2ЦڭYtTCRgB NX|D8Έ fP'l;/{E,Ѿ &OiZ΅` '7]GiuJDtRR)H6Ί@df c:0A Qc*{GMwȐ !5JL|A&F@q("sLS5klKgșT> 7=j#аDi"ӘLkկW-ԏD|2kx2PQ.j:iHĽ){H6z2y}EoQ6%h,82= 4G쌍b`md÷98$dkrE#n1& .?{W֑ ;IS/Xf18/ѫX"e.ATCQG-5yHE^u H% ?`mkMcv+/{?~@R$TP!G }@P@ 1=񲡖.tW{aia]Q?ჶìv/c oCyNcow7EvQ,e`GB{AL8Cѓ%P ,xXK s~JvCrg9մdJw,xa;zն(]Zc&xa x S׬BJ1: 4D.ȄJT<03@(4&U{1twA-Q* Ǜpyvس3_Ɠx޻^@9F?{$D tpu8&=*#&b^TSC7~i5jרPj h>qq<|tLAl0*P)ʍdNDX<6aiL83F`Wʆ(̹XE вZ.g [4ra%$Я 6Dv|磱|p}D/KM."j6e˻BjDRٷ"g<7} 329>9RAS5qf2X*R"3) `R-_D!91Gb (OEg1NKaԕIfvcC}e9$])okTޤ`C00 2-%s+s=*#\PqB(`"jAO.8' !Ds=I>KV#l(z!#h>w茸gg2ax)%3"~?ώT-KX`Z*',d>)Y (2V5'{yH7G"j`DR)S'4AldvhʱΈMhFof!N ^ȈܫSx5bF0CXAWEngGh&|q14w/?N%Hd2id(GZ ^s[LpO /qk? T i"OGbͶÏ\i#{FXR|EQ,F% DKު4BxL=;YpII]X쭥Z.-X|3wB.ݰ.Wη6ݼQz/zV$5s8x\)FK"ڢ0EY*7ISqٳ>G) p)DVI3p74eBӧfǮsӁKi!,+!rf}[~6n k"T[E"1zd}g(EqY*y"`*=(哀ʹ<:7zr5.q9pVE{qVZ2}VXLUZj^h]T.Rec5YKQK-RA"hE"՚j򬥇]xI6o7zQ~wrr_# 챝NȏG;;x)Lo6EjkhfVg(Yc_'YQtn94 K4&gnhh!l}nkXͻʄΫo~U[A-N&EbB N$ԴCX"ٕH6:K߅H6:Kwe6:K܅, Ϸk_2_i%ر܁7YQPCrʈ$xHCsθb} зC͚r:KXܵ|0x~c3 %CKS2L%L{DXQ7[*.Y]+\iApUžy;X7Ev~u_poeYa yqY B"Wf1u1=ygļNGo0(S=T.R s?$W&c6%F˹Vr(|^B֍+7 :tDt/ւZȲwㆲWse?92;u2* ոsGX9w@ZFPoçx8ʋq/5Hل۳ʄOGՑs~ә{=&dȁ Ae3fPi= ;yC:LR 9Ej 2F(1S^@ՎYf[,ؑEg8x_;كoaڛ9gbc@ ];U2 I!0!̓;Պ>QGB-mdMe(5hiDi/ Ɗ:"0W838qHŽ!(-BIo4%j,_sV넒j$ٻ'l!,}qdk^݂7HdZ@*:0Tj- r-%s麓wBJlqW܁tVXFiMfP;&xq֙fN_5f5}l؁u!s[GT}YoG}4|q|y8&‹lI<)%g]ML^W^6J%E.*wtfi/S\,*88Ǻqa;hMJtrߎ@<|A-{Yo흭l|F>+cK]>+gەen]YpveMƆ+W`yQfmm*%>,p ج 9mvPԾZWdN׺ّ[6FGW*a]@Ov+6L0|Y!ZO w~ ՚Fl:szju=m99wgk+eZlst?l4o~8= 66o[ o.,|ҙ/zV}y0#0:؁,S?|Q`d:; ob՞G6um]ߋk!q~ѐ'mt 2ɷ#Tuk˃iU.l`֭>nuۄB$18U >Q1/9wć&&&k«TދwQ}H5}\@!{l?T:6{myPtpY9pV9ށ=0<%tJA=K N:vb_㢔BYa5{> ^! !Zēq%ʴnYfL(LveZ=y9Sy٧4f ʽп_r.m_D3H|nKGa+H0U@hIJ?ta"'N&] -rxa*a3tؽ"eD Rqn;sTڣ>7>htNOmW8ɕ4T% !{Yq"#P2c/҈V^Ɛvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004151030615157214432017701 0ustar rootrootMar 20 08:57:12 crc systemd[1]: Starting Kubernetes Kubelet... Mar 20 08:57:12 crc restorecon[4693]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:12 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:13 crc restorecon[4693]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:57:13 crc restorecon[4693]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 20 08:57:13 crc kubenswrapper[4823]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 08:57:13 crc kubenswrapper[4823]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 20 08:57:13 crc kubenswrapper[4823]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 08:57:13 crc kubenswrapper[4823]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 08:57:13 crc kubenswrapper[4823]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 20 08:57:13 crc kubenswrapper[4823]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.795286 4823 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803649 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803707 4823 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803715 4823 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803722 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803728 4823 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803733 4823 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803744 4823 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803753 4823 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803758 4823 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803784 4823 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803790 4823 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803796 4823 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803804 4823 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803819 4823 feature_gate.go:330] unrecognized feature gate: Example Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803833 4823 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803840 4823 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803845 4823 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803851 4823 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803856 4823 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803861 4823 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803866 4823 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803871 4823 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803876 4823 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803882 4823 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803887 4823 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803892 4823 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803897 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803902 4823 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803907 4823 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803914 4823 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803919 4823 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803924 4823 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803929 4823 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803934 4823 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803941 4823 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803947 4823 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803952 4823 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803957 4823 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803962 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803968 4823 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803972 4823 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803978 4823 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803982 4823 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803987 4823 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803992 4823 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.803997 4823 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804002 4823 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804007 4823 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804014 4823 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804023 4823 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804029 4823 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804041 4823 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804051 4823 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804057 4823 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804068 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804074 4823 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804080 4823 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804086 4823 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804091 4823 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804097 4823 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804102 4823 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804108 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804114 4823 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804119 4823 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804125 4823 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804131 4823 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804137 4823 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804143 4823 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804148 4823 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804156 4823 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.804163 4823 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805198 4823 flags.go:64] FLAG: --address="0.0.0.0" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805230 4823 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805250 4823 flags.go:64] FLAG: --anonymous-auth="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805263 4823 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805275 4823 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805283 4823 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805295 4823 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805305 4823 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805313 4823 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805322 4823 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805331 4823 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805339 4823 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805347 4823 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805355 4823 flags.go:64] FLAG: --cgroup-root="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805364 4823 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805375 4823 flags.go:64] FLAG: --client-ca-file="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805382 4823 flags.go:64] FLAG: --cloud-config="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805390 4823 flags.go:64] FLAG: --cloud-provider="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805397 4823 flags.go:64] FLAG: --cluster-dns="[]" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805407 4823 flags.go:64] FLAG: --cluster-domain="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805414 4823 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805422 4823 flags.go:64] FLAG: --config-dir="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805429 4823 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805438 4823 flags.go:64] FLAG: --container-log-max-files="5" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805449 4823 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805457 4823 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805464 4823 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805472 4823 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805480 4823 flags.go:64] FLAG: --contention-profiling="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805487 4823 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805496 4823 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805503 4823 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805511 4823 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805521 4823 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805529 4823 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805537 4823 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805544 4823 flags.go:64] FLAG: --enable-load-reader="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805552 4823 flags.go:64] FLAG: --enable-server="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805559 4823 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805570 4823 flags.go:64] FLAG: --event-burst="100" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805578 4823 flags.go:64] FLAG: --event-qps="50" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805585 4823 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805592 4823 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805599 4823 flags.go:64] FLAG: --eviction-hard="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805609 4823 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805616 4823 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805624 4823 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805632 4823 flags.go:64] FLAG: --eviction-soft="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805640 4823 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805648 4823 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805660 4823 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805669 4823 flags.go:64] FLAG: --experimental-mounter-path="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805676 4823 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805683 4823 flags.go:64] FLAG: --fail-swap-on="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805690 4823 flags.go:64] FLAG: --feature-gates="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805700 4823 flags.go:64] FLAG: --file-check-frequency="20s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805708 4823 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805716 4823 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805723 4823 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805730 4823 flags.go:64] FLAG: --healthz-port="10248" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805738 4823 flags.go:64] FLAG: --help="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805752 4823 flags.go:64] FLAG: --hostname-override="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805761 4823 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805796 4823 flags.go:64] FLAG: --http-check-frequency="20s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805805 4823 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805812 4823 flags.go:64] FLAG: --image-credential-provider-config="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805853 4823 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805864 4823 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805872 4823 flags.go:64] FLAG: --image-service-endpoint="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805879 4823 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805887 4823 flags.go:64] FLAG: --kube-api-burst="100" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805895 4823 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805904 4823 flags.go:64] FLAG: --kube-api-qps="50" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805912 4823 flags.go:64] FLAG: --kube-reserved="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805921 4823 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805929 4823 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805937 4823 flags.go:64] FLAG: --kubelet-cgroups="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805945 4823 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805954 4823 flags.go:64] FLAG: --lock-file="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805963 4823 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805970 4823 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805978 4823 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805992 4823 flags.go:64] FLAG: --log-json-split-stream="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.805999 4823 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806007 4823 flags.go:64] FLAG: --log-text-split-stream="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806014 4823 flags.go:64] FLAG: --logging-format="text" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806024 4823 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806032 4823 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806040 4823 flags.go:64] FLAG: --manifest-url="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806047 4823 flags.go:64] FLAG: --manifest-url-header="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806058 4823 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806066 4823 flags.go:64] FLAG: --max-open-files="1000000" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806080 4823 flags.go:64] FLAG: --max-pods="110" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806088 4823 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806096 4823 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806103 4823 flags.go:64] FLAG: --memory-manager-policy="None" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806110 4823 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806118 4823 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806126 4823 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806134 4823 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806154 4823 flags.go:64] FLAG: --node-status-max-images="50" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806162 4823 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806169 4823 flags.go:64] FLAG: --oom-score-adj="-999" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806176 4823 flags.go:64] FLAG: --pod-cidr="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806183 4823 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806197 4823 flags.go:64] FLAG: --pod-manifest-path="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806205 4823 flags.go:64] FLAG: --pod-max-pids="-1" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806212 4823 flags.go:64] FLAG: --pods-per-core="0" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806220 4823 flags.go:64] FLAG: --port="10250" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806228 4823 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806236 4823 flags.go:64] FLAG: --provider-id="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806243 4823 flags.go:64] FLAG: --qos-reserved="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806251 4823 flags.go:64] FLAG: --read-only-port="10255" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806260 4823 flags.go:64] FLAG: --register-node="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806268 4823 flags.go:64] FLAG: --register-schedulable="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806276 4823 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806290 4823 flags.go:64] FLAG: --registry-burst="10" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806298 4823 flags.go:64] FLAG: --registry-qps="5" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806305 4823 flags.go:64] FLAG: --reserved-cpus="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806313 4823 flags.go:64] FLAG: --reserved-memory="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806324 4823 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806332 4823 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806341 4823 flags.go:64] FLAG: --rotate-certificates="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806349 4823 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806357 4823 flags.go:64] FLAG: --runonce="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806365 4823 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806373 4823 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806381 4823 flags.go:64] FLAG: --seccomp-default="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806388 4823 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806396 4823 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806404 4823 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806412 4823 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806420 4823 flags.go:64] FLAG: --storage-driver-password="root" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806427 4823 flags.go:64] FLAG: --storage-driver-secure="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806435 4823 flags.go:64] FLAG: --storage-driver-table="stats" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806442 4823 flags.go:64] FLAG: --storage-driver-user="root" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806449 4823 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806458 4823 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806466 4823 flags.go:64] FLAG: --system-cgroups="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806473 4823 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806486 4823 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806494 4823 flags.go:64] FLAG: --tls-cert-file="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806501 4823 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806511 4823 flags.go:64] FLAG: --tls-min-version="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806520 4823 flags.go:64] FLAG: --tls-private-key-file="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806528 4823 flags.go:64] FLAG: --topology-manager-policy="none" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806535 4823 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806545 4823 flags.go:64] FLAG: --topology-manager-scope="container" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806553 4823 flags.go:64] FLAG: --v="2" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806564 4823 flags.go:64] FLAG: --version="false" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806573 4823 flags.go:64] FLAG: --vmodule="" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806589 4823 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.806595 4823 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806754 4823 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806760 4823 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806788 4823 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806796 4823 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806802 4823 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806809 4823 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806816 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806822 4823 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806828 4823 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806834 4823 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806839 4823 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806844 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806850 4823 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806855 4823 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806861 4823 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806867 4823 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806872 4823 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806877 4823 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806883 4823 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806889 4823 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806895 4823 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806901 4823 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806907 4823 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806924 4823 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806930 4823 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806935 4823 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806940 4823 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806945 4823 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806950 4823 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806954 4823 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806960 4823 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806966 4823 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806971 4823 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806980 4823 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806988 4823 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.806994 4823 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807001 4823 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807007 4823 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807013 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807019 4823 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807025 4823 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807033 4823 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807040 4823 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807045 4823 feature_gate.go:330] unrecognized feature gate: Example Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807052 4823 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807058 4823 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807065 4823 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807071 4823 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807077 4823 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807085 4823 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807093 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807101 4823 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807108 4823 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807115 4823 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807122 4823 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807134 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807141 4823 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807148 4823 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807154 4823 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807160 4823 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807167 4823 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807173 4823 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807179 4823 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807185 4823 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807196 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807204 4823 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807212 4823 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807218 4823 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807224 4823 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807230 4823 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.807236 4823 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.807248 4823 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.855135 4823 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.855216 4823 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855364 4823 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855376 4823 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855382 4823 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855387 4823 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855392 4823 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855396 4823 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855402 4823 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855408 4823 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855413 4823 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855417 4823 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855422 4823 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855427 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855432 4823 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855437 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855442 4823 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855448 4823 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855453 4823 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855459 4823 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855465 4823 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855473 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855480 4823 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855486 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855493 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855500 4823 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855506 4823 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855511 4823 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855516 4823 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855522 4823 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855529 4823 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855533 4823 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855538 4823 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855543 4823 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855548 4823 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855554 4823 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855559 4823 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855564 4823 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855569 4823 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855574 4823 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855579 4823 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855585 4823 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855591 4823 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855596 4823 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855601 4823 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855608 4823 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855618 4823 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855625 4823 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855630 4823 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855637 4823 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855642 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855648 4823 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855654 4823 feature_gate.go:330] unrecognized feature gate: Example Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855659 4823 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855664 4823 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855670 4823 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855676 4823 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855682 4823 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855688 4823 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855693 4823 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855699 4823 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855705 4823 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855710 4823 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855716 4823 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855721 4823 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855726 4823 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855730 4823 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855735 4823 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855741 4823 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855747 4823 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855755 4823 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855760 4823 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855792 4823 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.855800 4823 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.855992 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856003 4823 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856009 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856018 4823 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856024 4823 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856029 4823 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856035 4823 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856040 4823 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856046 4823 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856051 4823 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856056 4823 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856063 4823 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856067 4823 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856075 4823 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856082 4823 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856088 4823 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856093 4823 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856098 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856103 4823 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856109 4823 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856115 4823 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856122 4823 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856127 4823 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856132 4823 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856138 4823 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856143 4823 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856148 4823 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856153 4823 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856158 4823 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856164 4823 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856169 4823 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856176 4823 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856182 4823 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856189 4823 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856195 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856201 4823 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856207 4823 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856213 4823 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856219 4823 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856224 4823 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856230 4823 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856234 4823 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856239 4823 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856244 4823 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856248 4823 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856253 4823 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856259 4823 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856263 4823 feature_gate.go:330] unrecognized feature gate: Example Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856268 4823 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856274 4823 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856280 4823 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856285 4823 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856290 4823 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856295 4823 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856300 4823 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856305 4823 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856310 4823 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856315 4823 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856319 4823 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856324 4823 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856330 4823 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856335 4823 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856343 4823 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856348 4823 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856353 4823 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856360 4823 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856364 4823 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856369 4823 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856374 4823 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856379 4823 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 08:57:13 crc kubenswrapper[4823]: W0320 08:57:13.856383 4823 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.856390 4823 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.858152 4823 server.go:940] "Client rotation is on, will bootstrap in background" Mar 20 08:57:13 crc kubenswrapper[4823]: E0320 08:57:13.862205 4823 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.867615 4823 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.867914 4823 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.869645 4823 server.go:997] "Starting client certificate rotation" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.869701 4823 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.869891 4823 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.892175 4823 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.895387 4823 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 20 08:57:13 crc kubenswrapper[4823]: E0320 08:57:13.895520 4823 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.914106 4823 log.go:25] "Validated CRI v1 runtime API" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.950282 4823 log.go:25] "Validated CRI v1 image API" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.953249 4823 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.961389 4823 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-20-08-51-41-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.961443 4823 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.991032 4823 manager.go:217] Machine: {Timestamp:2026-03-20 08:57:13.986338127 +0000 UTC m=+0.672234551 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:62ada254-c183-4bbc-aaa5-3bc10b9c932c BootID:8f9fe74c-818f-44ae-ba11-a9422f2d82f7 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:07:0a:19 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:07:0a:19 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:22:02:ea Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:84:ac:01 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:9a:a5:41 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:82:b4:31 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:7e:13:0e:e7:ca:80 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fe:7c:cd:78:d0:c7 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.991339 4823 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.991519 4823 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.992696 4823 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.992959 4823 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.993000 4823 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.993301 4823 topology_manager.go:138] "Creating topology manager with none policy" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.993318 4823 container_manager_linux.go:303] "Creating device plugin manager" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.993809 4823 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.993852 4823 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.994069 4823 state_mem.go:36] "Initialized new in-memory state store" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.994196 4823 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.998125 4823 kubelet.go:418] "Attempting to sync node with API server" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.998155 4823 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.998194 4823 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.998217 4823 kubelet.go:324] "Adding apiserver pod source" Mar 20 08:57:13 crc kubenswrapper[4823]: I0320 08:57:13.998236 4823 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 20 08:57:14 crc kubenswrapper[4823]: W0320 08:57:14.002986 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.003113 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.003302 4823 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 20 08:57:14 crc kubenswrapper[4823]: W0320 08:57:14.003344 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.003509 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.005220 4823 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.006442 4823 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009526 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009560 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009571 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009582 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009599 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009611 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009621 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009636 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009648 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009658 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009672 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.009682 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.014547 4823 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.015230 4823 server.go:1280] "Started kubelet" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.017059 4823 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.017053 4823 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.017687 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:14 crc systemd[1]: Started Kubernetes Kubelet. Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.018373 4823 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.019333 4823 server.go:460] "Adding debug handlers to kubelet server" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.021229 4823 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.021292 4823 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.021751 4823 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.021790 4823 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.021901 4823 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.021947 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:57:14 crc kubenswrapper[4823]: W0320 08:57:14.022669 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.022748 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.023081 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="200ms" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.023333 4823 factory.go:55] Registering systemd factory Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.023529 4823 factory.go:221] Registration of the systemd container factory successfully Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.029646 4823 factory.go:153] Registering CRI-O factory Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.024377 4823 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189e80e745126152 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.015187282 +0000 UTC m=+0.701083666,LastTimestamp:2026-03-20 08:57:14.015187282 +0000 UTC m=+0.701083666,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.029783 4823 factory.go:221] Registration of the crio container factory successfully Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.030340 4823 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.030572 4823 factory.go:103] Registering Raw factory Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.030619 4823 manager.go:1196] Started watching for new ooms in manager Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.034179 4823 manager.go:319] Starting recovery of all containers Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.041620 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.041759 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.041845 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.041874 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.041904 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.041933 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.041964 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.041992 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042030 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042058 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042086 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042113 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042139 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042177 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042208 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042237 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042264 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042297 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042328 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042372 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042462 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042493 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042524 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042553 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042580 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042612 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042650 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042687 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042716 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042753 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042827 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042865 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042899 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042940 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042970 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.042999 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043033 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043070 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043102 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043139 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043168 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043193 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043215 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043238 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043269 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043301 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043335 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043364 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043395 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043425 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043455 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043484 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043564 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043617 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043656 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043691 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043736 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043807 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043846 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043876 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043907 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043937 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.043974 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044003 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044036 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044065 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044097 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044126 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044158 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044190 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044219 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044249 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044279 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044305 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044334 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044364 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044401 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044466 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044504 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044537 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044566 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044598 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044629 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044662 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044692 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044737 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044810 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044853 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044885 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044919 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044953 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.044981 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.045018 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.045055 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.045095 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.045125 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.045158 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.045199 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.045233 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.045275 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.045307 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.045338 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.045381 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.047626 4823 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.047702 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.047761 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.047867 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.047912 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.047948 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.047982 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048017 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048051 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048081 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048111 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048145 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048175 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048204 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048234 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048261 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048291 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048321 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048350 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048382 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048423 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048578 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048613 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048650 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048691 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048724 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048754 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048840 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048876 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048944 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.048978 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049019 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049052 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049084 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049118 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049149 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049185 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049221 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049251 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049300 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049335 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049374 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049413 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049448 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049481 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049516 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049561 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049593 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.049624 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050182 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050223 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050260 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050292 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050325 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050370 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050404 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050435 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050470 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050501 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050534 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050563 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050590 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050618 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050648 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050679 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050712 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050752 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050824 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050855 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050889 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050923 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050960 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.050992 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051023 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051051 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051080 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051113 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051145 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051181 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051229 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051262 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051297 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051335 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051369 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051400 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051434 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051469 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051516 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051552 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051595 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051630 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051661 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051694 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051723 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051761 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051826 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051857 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051898 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051929 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051961 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.051995 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.052027 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.052048 4823 manager.go:324] Recovery completed Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.052060 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.052101 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.052134 4823 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.052160 4823 reconstruct.go:97] "Volume reconstruction finished" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.052182 4823 reconciler.go:26] "Reconciler: start to sync state" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.060616 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.062609 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.062662 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.062671 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.063647 4823 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.063668 4823 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.063686 4823 state_mem.go:36] "Initialized new in-memory state store" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.081805 4823 policy_none.go:49] "None policy: Start" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.089985 4823 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.090033 4823 state_mem.go:35] "Initializing new in-memory state store" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.093384 4823 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.095492 4823 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.095541 4823 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.095938 4823 kubelet.go:2335] "Starting kubelet main sync loop" Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.096004 4823 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 20 08:57:14 crc kubenswrapper[4823]: W0320 08:57:14.098106 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.098175 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.122183 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.140577 4823 manager.go:334] "Starting Device Plugin manager" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.140663 4823 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.140681 4823 server.go:79] "Starting device plugin registration server" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.141539 4823 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.141589 4823 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.141952 4823 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.142133 4823 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.142155 4823 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.152912 4823 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.196992 4823 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.197209 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.198994 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.199072 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.199087 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.199403 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.199786 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.199884 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.200687 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.200771 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.200785 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.200899 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.201088 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.201134 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.201687 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.201830 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.201880 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.202008 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.201941 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.202026 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.202074 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.202106 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.202132 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.202170 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.202438 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.202505 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.203164 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.203200 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.203218 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.203408 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.203430 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.203452 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.203461 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.203524 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.203575 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.205005 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.205042 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.205062 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.205087 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.205101 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.205067 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.205374 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.205412 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.206149 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.206191 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.206205 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.224195 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="400ms" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.242504 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.243529 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.243570 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.243585 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.243614 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.244014 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255132 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255179 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255293 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255389 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255418 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255501 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255559 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255641 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255699 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255737 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255787 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255819 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255846 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255873 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.255923 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.357569 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.357988 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.357880 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358149 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358212 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358287 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358315 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358390 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358462 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358499 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358643 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358692 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358685 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358566 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358719 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358805 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358800 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358850 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358880 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358929 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358957 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358968 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358981 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.359014 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.358997 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.359034 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.359033 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.359122 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.359041 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.359064 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.444780 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.446075 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.446099 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.446110 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.446133 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.446560 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.553231 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.574422 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.596263 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.623208 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.625459 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="800ms" Mar 20 08:57:14 crc kubenswrapper[4823]: W0320 08:57:14.629084 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-6e29df730e1f436fb9181a80f96dee9c699b7d3c0b838c57c1f05509e788a0e6 WatchSource:0}: Error finding container 6e29df730e1f436fb9181a80f96dee9c699b7d3c0b838c57c1f05509e788a0e6: Status 404 returned error can't find the container with id 6e29df730e1f436fb9181a80f96dee9c699b7d3c0b838c57c1f05509e788a0e6 Mar 20 08:57:14 crc kubenswrapper[4823]: W0320 08:57:14.629433 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-a6a1c7bee0636953de0fa9d8fc7892418ff33136a11380b1fa884fea94dc2a67 WatchSource:0}: Error finding container a6a1c7bee0636953de0fa9d8fc7892418ff33136a11380b1fa884fea94dc2a67: Status 404 returned error can't find the container with id a6a1c7bee0636953de0fa9d8fc7892418ff33136a11380b1fa884fea94dc2a67 Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.631794 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 20 08:57:14 crc kubenswrapper[4823]: W0320 08:57:14.646755 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-91d68ebf33f31862bf53447c2c01f50bd6e7c00ea89c4689cc6e000a696de3df WatchSource:0}: Error finding container 91d68ebf33f31862bf53447c2c01f50bd6e7c00ea89c4689cc6e000a696de3df: Status 404 returned error can't find the container with id 91d68ebf33f31862bf53447c2c01f50bd6e7c00ea89c4689cc6e000a696de3df Mar 20 08:57:14 crc kubenswrapper[4823]: W0320 08:57:14.648079 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-a9c3302cd51431898d949be08867c42b7378d8ac6132f3b896748d99611fe19a WatchSource:0}: Error finding container a9c3302cd51431898d949be08867c42b7378d8ac6132f3b896748d99611fe19a: Status 404 returned error can't find the container with id a9c3302cd51431898d949be08867c42b7378d8ac6132f3b896748d99611fe19a Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.846853 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.848394 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.848523 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.848585 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4823]: I0320 08:57:14.848637 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.849416 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Mar 20 08:57:14 crc kubenswrapper[4823]: E0320 08:57:14.855434 4823 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189e80e745126152 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.015187282 +0000 UTC m=+0.701083666,LastTimestamp:2026-03-20 08:57:14.015187282 +0000 UTC m=+0.701083666,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:15 crc kubenswrapper[4823]: I0320 08:57:15.019150 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:15 crc kubenswrapper[4823]: I0320 08:57:15.103019 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"91d68ebf33f31862bf53447c2c01f50bd6e7c00ea89c4689cc6e000a696de3df"} Mar 20 08:57:15 crc kubenswrapper[4823]: I0320 08:57:15.104563 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a2946fe69883fec101ba4a65adffa66dc4cfde0b1206a498ddf688dba2dfba65"} Mar 20 08:57:15 crc kubenswrapper[4823]: I0320 08:57:15.106036 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a6a1c7bee0636953de0fa9d8fc7892418ff33136a11380b1fa884fea94dc2a67"} Mar 20 08:57:15 crc kubenswrapper[4823]: I0320 08:57:15.108017 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6e29df730e1f436fb9181a80f96dee9c699b7d3c0b838c57c1f05509e788a0e6"} Mar 20 08:57:15 crc kubenswrapper[4823]: I0320 08:57:15.109314 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a9c3302cd51431898d949be08867c42b7378d8ac6132f3b896748d99611fe19a"} Mar 20 08:57:15 crc kubenswrapper[4823]: W0320 08:57:15.180476 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:15 crc kubenswrapper[4823]: E0320 08:57:15.181206 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:15 crc kubenswrapper[4823]: W0320 08:57:15.314714 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:15 crc kubenswrapper[4823]: E0320 08:57:15.314897 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:15 crc kubenswrapper[4823]: W0320 08:57:15.351119 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:15 crc kubenswrapper[4823]: E0320 08:57:15.351204 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:15 crc kubenswrapper[4823]: E0320 08:57:15.426571 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="1.6s" Mar 20 08:57:15 crc kubenswrapper[4823]: W0320 08:57:15.451598 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:15 crc kubenswrapper[4823]: E0320 08:57:15.451807 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:15 crc kubenswrapper[4823]: I0320 08:57:15.649922 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:15 crc kubenswrapper[4823]: I0320 08:57:15.652334 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:15 crc kubenswrapper[4823]: I0320 08:57:15.652375 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:15 crc kubenswrapper[4823]: I0320 08:57:15.652390 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:15 crc kubenswrapper[4823]: I0320 08:57:15.652421 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:15 crc kubenswrapper[4823]: E0320 08:57:15.653152 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.019647 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.082096 4823 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 08:57:16 crc kubenswrapper[4823]: E0320 08:57:16.083083 4823 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.119870 4823 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c" exitCode=0 Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.120004 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c"} Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.120058 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.121386 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.121435 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.121454 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.122719 4823 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7" exitCode=0 Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.122789 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7"} Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.122892 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.123906 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.123939 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.123948 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.124634 4823 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38" exitCode=0 Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.124713 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.124722 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38"} Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.126653 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.126682 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.126695 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.127995 4823 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5" exitCode=0 Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.128093 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.128091 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5"} Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.128863 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.128894 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.128906 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.131395 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519"} Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.131427 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182"} Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.131458 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc"} Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.131465 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.131470 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483"} Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.132130 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.132148 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.132156 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.132491 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.135941 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.135985 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.136003 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:16 crc kubenswrapper[4823]: I0320 08:57:16.272178 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.019154 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:17 crc kubenswrapper[4823]: E0320 08:57:17.028245 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="3.2s" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.140187 4823 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5" exitCode=0 Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.140270 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5"} Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.140431 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.141637 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.141676 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.141692 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.143612 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.144229 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1"} Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.145003 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.145031 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.145042 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.147818 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b1bdf60e1820b32bf856c731144b58c5a133edb212cd21952f53c679eef58699"} Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.147851 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ca0df1d0295ef37be29121ac847de5209abb269b53439eb479ffb9b82dad1a6a"} Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.147862 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.147863 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"61e8552c1d3d581d0f496b5f2cc3da10b7112c980788babf34c126bc9d3bc624"} Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.148844 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.148877 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.148888 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.152257 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8"} Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.152297 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e"} Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.152308 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89"} Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.152318 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a"} Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.152352 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.153212 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.153245 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.153258 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.253701 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.257172 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.257222 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.257236 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:17 crc kubenswrapper[4823]: I0320 08:57:17.257291 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:17 crc kubenswrapper[4823]: E0320 08:57:17.258123 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Mar 20 08:57:17 crc kubenswrapper[4823]: W0320 08:57:17.296373 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:17 crc kubenswrapper[4823]: E0320 08:57:17.296456 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:17 crc kubenswrapper[4823]: W0320 08:57:17.374747 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:17 crc kubenswrapper[4823]: E0320 08:57:17.375203 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:17 crc kubenswrapper[4823]: W0320 08:57:17.521803 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 20 08:57:17 crc kubenswrapper[4823]: E0320 08:57:17.521907 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.158163 4823 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e" exitCode=0 Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.158289 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e"} Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.158372 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.160113 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.160175 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.160195 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.163783 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e7f195b67056072d5964e394ae0a4034b76fe79c6ea6c8f66faf300a0a07bd32"} Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.163856 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.163839 4823 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.163901 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.163940 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.163920 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165439 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165513 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165526 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165687 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165719 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165761 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165781 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165805 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165722 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165848 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165863 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:18 crc kubenswrapper[4823]: I0320 08:57:18.165893 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.131057 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.171038 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4"} Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.171108 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10"} Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.171130 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc"} Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.171147 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2"} Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.171158 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d"} Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.171172 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.171172 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.172380 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.172413 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.172426 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.172408 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.172455 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.172467 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:19 crc kubenswrapper[4823]: I0320 08:57:19.383603 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.173543 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.173670 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.174938 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.174972 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.174985 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.175101 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.175128 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.175138 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.325295 4823 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.459275 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.461083 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.461139 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.461155 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.461188 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:20 crc kubenswrapper[4823]: I0320 08:57:20.819033 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.176912 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.178385 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.178441 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.178457 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.572478 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.572665 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.574627 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.574664 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.574677 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.581596 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.821815 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.822311 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.824204 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.824274 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:21 crc kubenswrapper[4823]: I0320 08:57:21.824298 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:22 crc kubenswrapper[4823]: I0320 08:57:22.179397 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:22 crc kubenswrapper[4823]: I0320 08:57:22.179468 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:22 crc kubenswrapper[4823]: I0320 08:57:22.180706 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:22 crc kubenswrapper[4823]: I0320 08:57:22.180801 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:22 crc kubenswrapper[4823]: I0320 08:57:22.180715 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:22 crc kubenswrapper[4823]: I0320 08:57:22.180875 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:22 crc kubenswrapper[4823]: I0320 08:57:22.180897 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:22 crc kubenswrapper[4823]: I0320 08:57:22.180829 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:22 crc kubenswrapper[4823]: I0320 08:57:22.222393 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:23 crc kubenswrapper[4823]: I0320 08:57:23.182671 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:23 crc kubenswrapper[4823]: I0320 08:57:23.184231 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:23 crc kubenswrapper[4823]: I0320 08:57:23.184292 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:23 crc kubenswrapper[4823]: I0320 08:57:23.184317 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.081040 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.081321 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.082719 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.082777 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.082790 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:24 crc kubenswrapper[4823]: E0320 08:57:24.153114 4823 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.162840 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.186001 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.187013 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.187081 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.187091 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.959597 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.959890 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.961241 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.961283 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:24 crc kubenswrapper[4823]: I0320 08:57:24.961294 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:26 crc kubenswrapper[4823]: I0320 08:57:26.278352 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:26 crc kubenswrapper[4823]: I0320 08:57:26.278493 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:26 crc kubenswrapper[4823]: I0320 08:57:26.279456 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:26 crc kubenswrapper[4823]: I0320 08:57:26.279494 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:26 crc kubenswrapper[4823]: I0320 08:57:26.279508 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:27 crc kubenswrapper[4823]: I0320 08:57:27.163381 4823 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 08:57:27 crc kubenswrapper[4823]: I0320 08:57:27.163483 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 08:57:28 crc kubenswrapper[4823]: W0320 08:57:28.000424 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 20 08:57:28 crc kubenswrapper[4823]: I0320 08:57:28.000637 4823 trace.go:236] Trace[765395001]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Mar-2026 08:57:17.998) (total time: 10002ms): Mar 20 08:57:28 crc kubenswrapper[4823]: Trace[765395001]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:57:28.000) Mar 20 08:57:28 crc kubenswrapper[4823]: Trace[765395001]: [10.002082877s] [10.002082877s] END Mar 20 08:57:28 crc kubenswrapper[4823]: E0320 08:57:28.000697 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 20 08:57:28 crc kubenswrapper[4823]: I0320 08:57:28.020460 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 20 08:57:28 crc kubenswrapper[4823]: E0320 08:57:28.177520 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 20 08:57:28 crc kubenswrapper[4823]: E0320 08:57:28.179228 4823 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:28 crc kubenswrapper[4823]: E0320 08:57:28.186505 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z" node="crc" Mar 20 08:57:28 crc kubenswrapper[4823]: W0320 08:57:28.188618 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z Mar 20 08:57:28 crc kubenswrapper[4823]: E0320 08:57:28.188702 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:28 crc kubenswrapper[4823]: W0320 08:57:28.191871 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z Mar 20 08:57:28 crc kubenswrapper[4823]: E0320 08:57:28.191951 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:28 crc kubenswrapper[4823]: I0320 08:57:28.192699 4823 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 20 08:57:28 crc kubenswrapper[4823]: I0320 08:57:28.192772 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 20 08:57:28 crc kubenswrapper[4823]: W0320 08:57:28.198494 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z Mar 20 08:57:28 crc kubenswrapper[4823]: E0320 08:57:28.198586 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:28 crc kubenswrapper[4823]: E0320 08:57:28.201415 4823 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189e80e745126152 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.015187282 +0000 UTC m=+0.701083666,LastTimestamp:2026-03-20 08:57:14.015187282 +0000 UTC m=+0.701083666,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:28 crc kubenswrapper[4823]: I0320 08:57:28.201510 4823 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 20 08:57:28 crc kubenswrapper[4823]: I0320 08:57:28.201577 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 20 08:57:29 crc kubenswrapper[4823]: I0320 08:57:29.022059 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:29Z is after 2026-02-23T05:33:13Z Mar 20 08:57:29 crc kubenswrapper[4823]: I0320 08:57:29.132355 4823 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 20 08:57:29 crc kubenswrapper[4823]: I0320 08:57:29.132431 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 20 08:57:29 crc kubenswrapper[4823]: I0320 08:57:29.198905 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 20 08:57:29 crc kubenswrapper[4823]: I0320 08:57:29.201401 4823 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e7f195b67056072d5964e394ae0a4034b76fe79c6ea6c8f66faf300a0a07bd32" exitCode=255 Mar 20 08:57:29 crc kubenswrapper[4823]: I0320 08:57:29.201452 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e7f195b67056072d5964e394ae0a4034b76fe79c6ea6c8f66faf300a0a07bd32"} Mar 20 08:57:29 crc kubenswrapper[4823]: I0320 08:57:29.201612 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:29 crc kubenswrapper[4823]: I0320 08:57:29.202538 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:29 crc kubenswrapper[4823]: I0320 08:57:29.202594 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:29 crc kubenswrapper[4823]: I0320 08:57:29.202612 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:29 crc kubenswrapper[4823]: I0320 08:57:29.203573 4823 scope.go:117] "RemoveContainer" containerID="e7f195b67056072d5964e394ae0a4034b76fe79c6ea6c8f66faf300a0a07bd32" Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.024244 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:30Z is after 2026-02-23T05:33:13Z Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.207885 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.208564 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.210176 4823 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3fda6b6e4da6f6cff2e64233e6e8daad78457d875abc4e1b379fcd8e672f642b" exitCode=255 Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.210232 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3fda6b6e4da6f6cff2e64233e6e8daad78457d875abc4e1b379fcd8e672f642b"} Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.210290 4823 scope.go:117] "RemoveContainer" containerID="e7f195b67056072d5964e394ae0a4034b76fe79c6ea6c8f66faf300a0a07bd32" Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.210624 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.212552 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.212590 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.212600 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.213210 4823 scope.go:117] "RemoveContainer" containerID="3fda6b6e4da6f6cff2e64233e6e8daad78457d875abc4e1b379fcd8e672f642b" Mar 20 08:57:30 crc kubenswrapper[4823]: E0320 08:57:30.213428 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:30 crc kubenswrapper[4823]: I0320 08:57:30.827778 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:31 crc kubenswrapper[4823]: I0320 08:57:31.022012 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:31Z is after 2026-02-23T05:33:13Z Mar 20 08:57:31 crc kubenswrapper[4823]: I0320 08:57:31.215408 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 20 08:57:31 crc kubenswrapper[4823]: I0320 08:57:31.217392 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:31 crc kubenswrapper[4823]: I0320 08:57:31.218848 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:31 crc kubenswrapper[4823]: I0320 08:57:31.218922 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:31 crc kubenswrapper[4823]: I0320 08:57:31.218936 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:31 crc kubenswrapper[4823]: I0320 08:57:31.219722 4823 scope.go:117] "RemoveContainer" containerID="3fda6b6e4da6f6cff2e64233e6e8daad78457d875abc4e1b379fcd8e672f642b" Mar 20 08:57:31 crc kubenswrapper[4823]: E0320 08:57:31.220084 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:31 crc kubenswrapper[4823]: I0320 08:57:31.222661 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:32 crc kubenswrapper[4823]: I0320 08:57:32.021915 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:32Z is after 2026-02-23T05:33:13Z Mar 20 08:57:32 crc kubenswrapper[4823]: I0320 08:57:32.220145 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:32 crc kubenswrapper[4823]: I0320 08:57:32.221178 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:32 crc kubenswrapper[4823]: I0320 08:57:32.221230 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:32 crc kubenswrapper[4823]: I0320 08:57:32.221250 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:32 crc kubenswrapper[4823]: I0320 08:57:32.222101 4823 scope.go:117] "RemoveContainer" containerID="3fda6b6e4da6f6cff2e64233e6e8daad78457d875abc4e1b379fcd8e672f642b" Mar 20 08:57:32 crc kubenswrapper[4823]: E0320 08:57:32.222384 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:33 crc kubenswrapper[4823]: I0320 08:57:33.021861 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:33Z is after 2026-02-23T05:33:13Z Mar 20 08:57:33 crc kubenswrapper[4823]: I0320 08:57:33.642242 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:33 crc kubenswrapper[4823]: I0320 08:57:33.642448 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:33 crc kubenswrapper[4823]: I0320 08:57:33.643942 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:33 crc kubenswrapper[4823]: I0320 08:57:33.644031 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:33 crc kubenswrapper[4823]: I0320 08:57:33.644047 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:33 crc kubenswrapper[4823]: I0320 08:57:33.644993 4823 scope.go:117] "RemoveContainer" containerID="3fda6b6e4da6f6cff2e64233e6e8daad78457d875abc4e1b379fcd8e672f642b" Mar 20 08:57:33 crc kubenswrapper[4823]: E0320 08:57:33.645250 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:33 crc kubenswrapper[4823]: W0320 08:57:33.861257 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 20 08:57:33 crc kubenswrapper[4823]: E0320 08:57:33.861320 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 20 08:57:34 crc kubenswrapper[4823]: I0320 08:57:34.026379 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:34 crc kubenswrapper[4823]: E0320 08:57:34.153605 4823 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:57:34 crc kubenswrapper[4823]: E0320 08:57:34.586000 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:57:34 crc kubenswrapper[4823]: I0320 08:57:34.587035 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:34 crc kubenswrapper[4823]: I0320 08:57:34.588649 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:34 crc kubenswrapper[4823]: I0320 08:57:34.588701 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:34 crc kubenswrapper[4823]: I0320 08:57:34.588714 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:34 crc kubenswrapper[4823]: I0320 08:57:34.588772 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:34 crc kubenswrapper[4823]: E0320 08:57:34.596347 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:57:35 crc kubenswrapper[4823]: I0320 08:57:35.023663 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:35 crc kubenswrapper[4823]: I0320 08:57:35.023946 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 20 08:57:35 crc kubenswrapper[4823]: I0320 08:57:35.024145 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:35 crc kubenswrapper[4823]: I0320 08:57:35.025330 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:35 crc kubenswrapper[4823]: I0320 08:57:35.025406 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:35 crc kubenswrapper[4823]: I0320 08:57:35.025444 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:35 crc kubenswrapper[4823]: I0320 08:57:35.036320 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 20 08:57:35 crc kubenswrapper[4823]: I0320 08:57:35.226839 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:35 crc kubenswrapper[4823]: I0320 08:57:35.228145 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:35 crc kubenswrapper[4823]: I0320 08:57:35.228189 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:35 crc kubenswrapper[4823]: I0320 08:57:35.228204 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:36 crc kubenswrapper[4823]: I0320 08:57:36.024331 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:36 crc kubenswrapper[4823]: I0320 08:57:36.781995 4823 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 08:57:36 crc kubenswrapper[4823]: I0320 08:57:36.801179 4823 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 20 08:57:37 crc kubenswrapper[4823]: I0320 08:57:37.024296 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:37 crc kubenswrapper[4823]: I0320 08:57:37.164386 4823 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 08:57:37 crc kubenswrapper[4823]: I0320 08:57:37.164541 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 08:57:38 crc kubenswrapper[4823]: I0320 08:57:38.023297 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.208723 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e745126152 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.015187282 +0000 UTC m=+0.701083666,LastTimestamp:2026-03-20 08:57:14.015187282 +0000 UTC m=+0.701083666,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.217413 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6aa3f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062654015 +0000 UTC m=+0.748550389,LastTimestamp:2026-03-20 08:57:14.062654015 +0000 UTC m=+0.748550389,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.223060 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6e44b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062668875 +0000 UTC m=+0.748565249,LastTimestamp:2026-03-20 08:57:14.062668875 +0000 UTC m=+0.748565249,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.229276 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e707d8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062677976 +0000 UTC m=+0.748574350,LastTimestamp:2026-03-20 08:57:14.062677976 +0000 UTC m=+0.748574350,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.233354 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e74cdc36a0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.145855136 +0000 UTC m=+0.831751510,LastTimestamp:2026-03-20 08:57:14.145855136 +0000 UTC m=+0.831751510,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.236341 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6aa3f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6aa3f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062654015 +0000 UTC m=+0.748550389,LastTimestamp:2026-03-20 08:57:14.199028056 +0000 UTC m=+0.884924430,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.239672 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6e44b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6e44b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062668875 +0000 UTC m=+0.748565249,LastTimestamp:2026-03-20 08:57:14.199082398 +0000 UTC m=+0.884978772,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.244185 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e707d8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e707d8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062677976 +0000 UTC m=+0.748574350,LastTimestamp:2026-03-20 08:57:14.199093308 +0000 UTC m=+0.884989682,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.248855 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6aa3f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6aa3f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062654015 +0000 UTC m=+0.748550389,LastTimestamp:2026-03-20 08:57:14.200707473 +0000 UTC m=+0.886603847,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.253008 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6e44b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6e44b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062668875 +0000 UTC m=+0.748565249,LastTimestamp:2026-03-20 08:57:14.200779695 +0000 UTC m=+0.886676079,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.257541 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e707d8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e707d8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062677976 +0000 UTC m=+0.748574350,LastTimestamp:2026-03-20 08:57:14.200790205 +0000 UTC m=+0.886686579,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.261562 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6aa3f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6aa3f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062654015 +0000 UTC m=+0.748550389,LastTimestamp:2026-03-20 08:57:14.201810046 +0000 UTC m=+0.887706430,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.266345 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6e44b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6e44b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062668875 +0000 UTC m=+0.748565249,LastTimestamp:2026-03-20 08:57:14.201928101 +0000 UTC m=+0.887824475,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.270994 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6aa3f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6aa3f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062654015 +0000 UTC m=+0.748550389,LastTimestamp:2026-03-20 08:57:14.201969622 +0000 UTC m=+0.887865996,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.275305 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6e44b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6e44b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062668875 +0000 UTC m=+0.748565249,LastTimestamp:2026-03-20 08:57:14.202020424 +0000 UTC m=+0.887916798,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.279732 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e707d8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e707d8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062677976 +0000 UTC m=+0.748574350,LastTimestamp:2026-03-20 08:57:14.202035215 +0000 UTC m=+0.887931599,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.284241 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e707d8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e707d8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062677976 +0000 UTC m=+0.748574350,LastTimestamp:2026-03-20 08:57:14.202078057 +0000 UTC m=+0.887974441,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.290543 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6aa3f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6aa3f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062654015 +0000 UTC m=+0.748550389,LastTimestamp:2026-03-20 08:57:14.202093847 +0000 UTC m=+0.887990231,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.295092 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6e44b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6e44b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062668875 +0000 UTC m=+0.748565249,LastTimestamp:2026-03-20 08:57:14.202121729 +0000 UTC m=+0.888018113,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.300416 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e707d8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e707d8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062677976 +0000 UTC m=+0.748574350,LastTimestamp:2026-03-20 08:57:14.202140909 +0000 UTC m=+0.888037293,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: W0320 08:57:38.305449 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.305533 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.305489 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6aa3f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6aa3f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062654015 +0000 UTC m=+0.748550389,LastTimestamp:2026-03-20 08:57:14.203183891 +0000 UTC m=+0.889080275,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.310523 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6e44b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6e44b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062668875 +0000 UTC m=+0.748565249,LastTimestamp:2026-03-20 08:57:14.203207802 +0000 UTC m=+0.889104196,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.316627 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e707d8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e707d8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062677976 +0000 UTC m=+0.748574350,LastTimestamp:2026-03-20 08:57:14.203225823 +0000 UTC m=+0.889122217,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.321548 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6aa3f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6aa3f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062654015 +0000 UTC m=+0.748550389,LastTimestamp:2026-03-20 08:57:14.203445291 +0000 UTC m=+0.889341665,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.325709 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e747e6e44b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e747e6e44b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.062668875 +0000 UTC m=+0.748565249,LastTimestamp:2026-03-20 08:57:14.203457552 +0000 UTC m=+0.889353926,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.333527 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e76a6dcf34 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.64193618 +0000 UTC m=+1.327832574,LastTimestamp:2026-03-20 08:57:14.64193618 +0000 UTC m=+1.327832574,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.339290 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e76a6e60cb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.641973451 +0000 UTC m=+1.327869825,LastTimestamp:2026-03-20 08:57:14.641973451 +0000 UTC m=+1.327869825,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.342115 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e76a7edcff openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.643053823 +0000 UTC m=+1.328950207,LastTimestamp:2026-03-20 08:57:14.643053823 +0000 UTC m=+1.328950207,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.343841 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e76afbee49 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.651250249 +0000 UTC m=+1.337146623,LastTimestamp:2026-03-20 08:57:14.651250249 +0000 UTC m=+1.337146623,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.348315 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e76b20a915 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.653657365 +0000 UTC m=+1.339553739,LastTimestamp:2026-03-20 08:57:14.653657365 +0000 UTC m=+1.339553739,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.353762 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e78c675624 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.211937316 +0000 UTC m=+1.897833690,LastTimestamp:2026-03-20 08:57:15.211937316 +0000 UTC m=+1.897833690,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.359151 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e78c7bb419 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.213272089 +0000 UTC m=+1.899168453,LastTimestamp:2026-03-20 08:57:15.213272089 +0000 UTC m=+1.899168453,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.365003 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e78c7e0d19 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.213425945 +0000 UTC m=+1.899322329,LastTimestamp:2026-03-20 08:57:15.213425945 +0000 UTC m=+1.899322329,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.371869 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e78c831ff1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.213758449 +0000 UTC m=+1.899654823,LastTimestamp:2026-03-20 08:57:15.213758449 +0000 UTC m=+1.899654823,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.377590 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e78c8ac96c openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.214260588 +0000 UTC m=+1.900156962,LastTimestamp:2026-03-20 08:57:15.214260588 +0000 UTC m=+1.900156962,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.383431 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e78d7bf5ec openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.230066156 +0000 UTC m=+1.915962560,LastTimestamp:2026-03-20 08:57:15.230066156 +0000 UTC m=+1.915962560,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.389418 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e78d7c726b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.230098027 +0000 UTC m=+1.915994401,LastTimestamp:2026-03-20 08:57:15.230098027 +0000 UTC m=+1.915994401,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.394401 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e78d9b01b2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.232100786 +0000 UTC m=+1.917997160,LastTimestamp:2026-03-20 08:57:15.232100786 +0000 UTC m=+1.917997160,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.400962 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e78dadeade openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.233340126 +0000 UTC m=+1.919236500,LastTimestamp:2026-03-20 08:57:15.233340126 +0000 UTC m=+1.919236500,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.406706 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e78dc1ae21 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.234635297 +0000 UTC m=+1.920531661,LastTimestamp:2026-03-20 08:57:15.234635297 +0000 UTC m=+1.920531661,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.412359 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e78dc4a966 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.234830694 +0000 UTC m=+1.920727068,LastTimestamp:2026-03-20 08:57:15.234830694 +0000 UTC m=+1.920727068,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.417551 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e79f9a2154 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.534033236 +0000 UTC m=+2.219929610,LastTimestamp:2026-03-20 08:57:15.534033236 +0000 UTC m=+2.219929610,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.423325 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e7a074233e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.548320574 +0000 UTC m=+2.234216948,LastTimestamp:2026-03-20 08:57:15.548320574 +0000 UTC m=+2.234216948,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.431097 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e7a08a5472 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.549774962 +0000 UTC m=+2.235671336,LastTimestamp:2026-03-20 08:57:15.549774962 +0000 UTC m=+2.235671336,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.438237 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e7aba6a1d6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.736179158 +0000 UTC m=+2.422075532,LastTimestamp:2026-03-20 08:57:15.736179158 +0000 UTC m=+2.422075532,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.445010 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e7ac7a4d38 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.750051128 +0000 UTC m=+2.435947502,LastTimestamp:2026-03-20 08:57:15.750051128 +0000 UTC m=+2.435947502,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.448842 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e7ac8bf2d2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.751207634 +0000 UTC m=+2.437104008,LastTimestamp:2026-03-20 08:57:15.751207634 +0000 UTC m=+2.437104008,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.453372 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e7bd47b853 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.031948883 +0000 UTC m=+2.717845257,LastTimestamp:2026-03-20 08:57:16.031948883 +0000 UTC m=+2.717845257,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.458154 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e7be74869f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.051662495 +0000 UTC m=+2.737558869,LastTimestamp:2026-03-20 08:57:16.051662495 +0000 UTC m=+2.737558869,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.462853 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e7c2b61989 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.123068809 +0000 UTC m=+2.808965183,LastTimestamp:2026-03-20 08:57:16.123068809 +0000 UTC m=+2.808965183,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.470543 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e7c2dbd081 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.125540481 +0000 UTC m=+2.811436855,LastTimestamp:2026-03-20 08:57:16.125540481 +0000 UTC m=+2.811436855,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.477309 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e7c308151d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.128441629 +0000 UTC m=+2.814338003,LastTimestamp:2026-03-20 08:57:16.128441629 +0000 UTC m=+2.814338003,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.482688 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7c3429beb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.132277227 +0000 UTC m=+2.818173601,LastTimestamp:2026-03-20 08:57:16.132277227 +0000 UTC m=+2.818173601,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.491285 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e7d42dd62e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.416128558 +0000 UTC m=+3.102024932,LastTimestamp:2026-03-20 08:57:16.416128558 +0000 UTC m=+3.102024932,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.497882 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e7d45b0f8f openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.419092367 +0000 UTC m=+3.104988741,LastTimestamp:2026-03-20 08:57:16.419092367 +0000 UTC m=+3.104988741,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.502936 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e7d4631c41 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.419619905 +0000 UTC m=+3.105516269,LastTimestamp:2026-03-20 08:57:16.419619905 +0000 UTC m=+3.105516269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.507688 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7d467f15e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.419936606 +0000 UTC m=+3.105832990,LastTimestamp:2026-03-20 08:57:16.419936606 +0000 UTC m=+3.105832990,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.512137 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e7d50f2bbf openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.430896063 +0000 UTC m=+3.116792437,LastTimestamp:2026-03-20 08:57:16.430896063 +0000 UTC m=+3.116792437,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.519056 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e7d5230a48 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.432198216 +0000 UTC m=+3.118094620,LastTimestamp:2026-03-20 08:57:16.432198216 +0000 UTC m=+3.118094620,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.524113 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e7d609d546 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.447323462 +0000 UTC m=+3.133219836,LastTimestamp:2026-03-20 08:57:16.447323462 +0000 UTC m=+3.133219836,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.531409 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7d650b281 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.451967617 +0000 UTC m=+3.137864001,LastTimestamp:2026-03-20 08:57:16.451967617 +0000 UTC m=+3.137864001,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.536339 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7d66429e0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.45324336 +0000 UTC m=+3.139139734,LastTimestamp:2026-03-20 08:57:16.45324336 +0000 UTC m=+3.139139734,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.543282 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e7d6805446 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.455089222 +0000 UTC m=+3.140985596,LastTimestamp:2026-03-20 08:57:16.455089222 +0000 UTC m=+3.140985596,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.548650 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e7e21480d7 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.649349335 +0000 UTC m=+3.335245739,LastTimestamp:2026-03-20 08:57:16.649349335 +0000 UTC m=+3.335245739,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.551523 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e7e316f8b8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.666288312 +0000 UTC m=+3.352184696,LastTimestamp:2026-03-20 08:57:16.666288312 +0000 UTC m=+3.352184696,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.553765 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e7e338c26c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.668502636 +0000 UTC m=+3.354399020,LastTimestamp:2026-03-20 08:57:16.668502636 +0000 UTC m=+3.354399020,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.559349 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7e4fe2b40 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.69821728 +0000 UTC m=+3.384113654,LastTimestamp:2026-03-20 08:57:16.69821728 +0000 UTC m=+3.384113654,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.563322 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7e5e98231 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.713640497 +0000 UTC m=+3.399536871,LastTimestamp:2026-03-20 08:57:16.713640497 +0000 UTC m=+3.399536871,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.567104 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7e5f9f1b7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.714717623 +0000 UTC m=+3.400613997,LastTimestamp:2026-03-20 08:57:16.714717623 +0000 UTC m=+3.400613997,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.571881 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e7f0a54113 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.893716755 +0000 UTC m=+3.579613139,LastTimestamp:2026-03-20 08:57:16.893716755 +0000 UTC m=+3.579613139,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.577056 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e7f1b79e30 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.911697456 +0000 UTC m=+3.597593830,LastTimestamp:2026-03-20 08:57:16.911697456 +0000 UTC m=+3.597593830,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.582031 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7f3ec3ecb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.948700875 +0000 UTC m=+3.634597249,LastTimestamp:2026-03-20 08:57:16.948700875 +0000 UTC m=+3.634597249,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.586703 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7f4b13ba0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.961610656 +0000 UTC m=+3.647507030,LastTimestamp:2026-03-20 08:57:16.961610656 +0000 UTC m=+3.647507030,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.592453 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7f4c35647 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:16.962797127 +0000 UTC m=+3.648693501,LastTimestamp:2026-03-20 08:57:16.962797127 +0000 UTC m=+3.648693501,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.599070 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7fe7863f4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:17.125657588 +0000 UTC m=+3.811553962,LastTimestamp:2026-03-20 08:57:17.125657588 +0000 UTC m=+3.811553962,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.604294 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7ff65d6a2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:17.141218978 +0000 UTC m=+3.827115352,LastTimestamp:2026-03-20 08:57:17.141218978 +0000 UTC m=+3.827115352,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: W0320 08:57:38.613469 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.613525 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.613762 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7ff7e146e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:17.142807662 +0000 UTC m=+3.828704036,LastTimestamp:2026-03-20 08:57:17.142807662 +0000 UTC m=+3.828704036,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.620108 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e7ffa9f5b2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:17.145683378 +0000 UTC m=+3.831579752,LastTimestamp:2026-03-20 08:57:17.145683378 +0000 UTC m=+3.831579752,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.625325 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e80b6722aa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:17.34263057 +0000 UTC m=+4.028526944,LastTimestamp:2026-03-20 08:57:17.34263057 +0000 UTC m=+4.028526944,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.629814 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e80b7257b5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:17.343365045 +0000 UTC m=+4.029261409,LastTimestamp:2026-03-20 08:57:17.343365045 +0000 UTC m=+4.029261409,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.634872 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e80c448ce0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:17.357141216 +0000 UTC m=+4.043037590,LastTimestamp:2026-03-20 08:57:17.357141216 +0000 UTC m=+4.043037590,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.639473 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e80c6e0661 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:17.359859297 +0000 UTC m=+4.045755671,LastTimestamp:2026-03-20 08:57:17.359859297 +0000 UTC m=+4.045755671,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.644193 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e83c443778 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.16242572 +0000 UTC m=+4.848322094,LastTimestamp:2026-03-20 08:57:18.16242572 +0000 UTC m=+4.848322094,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.649976 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e846fa9e38 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.342151736 +0000 UTC m=+5.028048150,LastTimestamp:2026-03-20 08:57:18.342151736 +0000 UTC m=+5.028048150,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.654227 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e847962bac openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.352346028 +0000 UTC m=+5.038242412,LastTimestamp:2026-03-20 08:57:18.352346028 +0000 UTC m=+5.038242412,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.657972 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e847ae32f9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.353920761 +0000 UTC m=+5.039817135,LastTimestamp:2026-03-20 08:57:18.353920761 +0000 UTC m=+5.039817135,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.662872 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e851cd40e8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.523728104 +0000 UTC m=+5.209624478,LastTimestamp:2026-03-20 08:57:18.523728104 +0000 UTC m=+5.209624478,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.668507 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e852622bdd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.533487581 +0000 UTC m=+5.219383955,LastTimestamp:2026-03-20 08:57:18.533487581 +0000 UTC m=+5.219383955,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.673584 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e85272d501 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.534579457 +0000 UTC m=+5.220475821,LastTimestamp:2026-03-20 08:57:18.534579457 +0000 UTC m=+5.220475821,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.677700 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e85d83e3ea openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.720246762 +0000 UTC m=+5.406143126,LastTimestamp:2026-03-20 08:57:18.720246762 +0000 UTC m=+5.406143126,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.681943 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e85e287d7f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.731033983 +0000 UTC m=+5.416930357,LastTimestamp:2026-03-20 08:57:18.731033983 +0000 UTC m=+5.416930357,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.686186 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e85e3bfcae openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.732311726 +0000 UTC m=+5.418208100,LastTimestamp:2026-03-20 08:57:18.732311726 +0000 UTC m=+5.418208100,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.691328 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e868673cb7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.902918327 +0000 UTC m=+5.588814701,LastTimestamp:2026-03-20 08:57:18.902918327 +0000 UTC m=+5.588814701,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.695846 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e869126768 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.914135912 +0000 UTC m=+5.600032286,LastTimestamp:2026-03-20 08:57:18.914135912 +0000 UTC m=+5.600032286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.701359 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e869226fec openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:18.915186668 +0000 UTC m=+5.601083042,LastTimestamp:2026-03-20 08:57:18.915186668 +0000 UTC m=+5.601083042,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.705578 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e871441f64 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:19.051612004 +0000 UTC m=+5.737508378,LastTimestamp:2026-03-20 08:57:19.051612004 +0000 UTC m=+5.737508378,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.709226 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e871fe1b41 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:19.063800641 +0000 UTC m=+5.749697015,LastTimestamp:2026-03-20 08:57:19.063800641 +0000 UTC m=+5.749697015,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.716134 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 08:57:38 crc kubenswrapper[4823]: &Event{ObjectMeta:{kube-controller-manager-crc.189e80ea54c4f8cc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 08:57:38 crc kubenswrapper[4823]: body: Mar 20 08:57:38 crc kubenswrapper[4823]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:27.163451596 +0000 UTC m=+13.849347980,LastTimestamp:2026-03-20 08:57:27.163451596 +0000 UTC m=+13.849347980,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:38 crc kubenswrapper[4823]: > Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.720164 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80ea54c6479c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:27.163537308 +0000 UTC m=+13.849433702,LastTimestamp:2026-03-20 08:57:27.163537308 +0000 UTC m=+13.849433702,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.727416 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 20 08:57:38 crc kubenswrapper[4823]: &Event{ObjectMeta:{kube-apiserver-crc.189e80ea921e865a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 20 08:57:38 crc kubenswrapper[4823]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 20 08:57:38 crc kubenswrapper[4823]: Mar 20 08:57:38 crc kubenswrapper[4823]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:28.192730714 +0000 UTC m=+14.878627088,LastTimestamp:2026-03-20 08:57:28.192730714 +0000 UTC m=+14.878627088,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:38 crc kubenswrapper[4823]: > Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.731924 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80ea921f8931 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:28.192796977 +0000 UTC m=+14.878693351,LastTimestamp:2026-03-20 08:57:28.192796977 +0000 UTC m=+14.878693351,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.738537 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e80ea921e865a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 20 08:57:38 crc kubenswrapper[4823]: &Event{ObjectMeta:{kube-apiserver-crc.189e80ea921e865a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 20 08:57:38 crc kubenswrapper[4823]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 20 08:57:38 crc kubenswrapper[4823]: Mar 20 08:57:38 crc kubenswrapper[4823]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:28.192730714 +0000 UTC m=+14.878627088,LastTimestamp:2026-03-20 08:57:28.201558448 +0000 UTC m=+14.887454822,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:38 crc kubenswrapper[4823]: > Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.743228 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e80ea921f8931\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80ea921f8931 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:28.192796977 +0000 UTC m=+14.878693351,LastTimestamp:2026-03-20 08:57:28.201600759 +0000 UTC m=+14.887497133,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.749701 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 20 08:57:38 crc kubenswrapper[4823]: &Event{ObjectMeta:{kube-apiserver-crc.189e80eaca20ee3f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Mar 20 08:57:38 crc kubenswrapper[4823]: body: Mar 20 08:57:38 crc kubenswrapper[4823]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:29.132412479 +0000 UTC m=+15.818308873,LastTimestamp:2026-03-20 08:57:29.132412479 +0000 UTC m=+15.818308873,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:38 crc kubenswrapper[4823]: > Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.756396 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80eaca21b047 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:29.132462151 +0000 UTC m=+15.818358535,LastTimestamp:2026-03-20 08:57:29.132462151 +0000 UTC m=+15.818358535,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.759580 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e80e7ff7e146e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e7ff7e146e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:17.142807662 +0000 UTC m=+3.828704036,LastTimestamp:2026-03-20 08:57:29.205240523 +0000 UTC m=+15.891136937,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.762943 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80ea54c4f8cc\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 08:57:38 crc kubenswrapper[4823]: &Event{ObjectMeta:{kube-controller-manager-crc.189e80ea54c4f8cc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 08:57:38 crc kubenswrapper[4823]: body: Mar 20 08:57:38 crc kubenswrapper[4823]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:27.163451596 +0000 UTC m=+13.849347980,LastTimestamp:2026-03-20 08:57:37.16449591 +0000 UTC m=+23.850392324,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:38 crc kubenswrapper[4823]: > Mar 20 08:57:38 crc kubenswrapper[4823]: E0320 08:57:38.767639 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80ea54c6479c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80ea54c6479c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:27.163537308 +0000 UTC m=+13.849433702,LastTimestamp:2026-03-20 08:57:37.164607694 +0000 UTC m=+23.850504118,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:39 crc kubenswrapper[4823]: I0320 08:57:39.026799 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:39 crc kubenswrapper[4823]: W0320 08:57:39.071202 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 20 08:57:39 crc kubenswrapper[4823]: E0320 08:57:39.071306 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 20 08:57:39 crc kubenswrapper[4823]: I0320 08:57:39.131116 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:39 crc kubenswrapper[4823]: I0320 08:57:39.131489 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:39 crc kubenswrapper[4823]: I0320 08:57:39.133427 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:39 crc kubenswrapper[4823]: I0320 08:57:39.133594 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:39 crc kubenswrapper[4823]: I0320 08:57:39.133728 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:39 crc kubenswrapper[4823]: I0320 08:57:39.134550 4823 scope.go:117] "RemoveContainer" containerID="3fda6b6e4da6f6cff2e64233e6e8daad78457d875abc4e1b379fcd8e672f642b" Mar 20 08:57:39 crc kubenswrapper[4823]: E0320 08:57:39.134908 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:40 crc kubenswrapper[4823]: I0320 08:57:40.024736 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:41 crc kubenswrapper[4823]: I0320 08:57:41.025156 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:41 crc kubenswrapper[4823]: W0320 08:57:41.137208 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 20 08:57:41 crc kubenswrapper[4823]: E0320 08:57:41.137610 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 20 08:57:41 crc kubenswrapper[4823]: E0320 08:57:41.591699 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:57:41 crc kubenswrapper[4823]: I0320 08:57:41.596652 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:41 crc kubenswrapper[4823]: I0320 08:57:41.598212 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:41 crc kubenswrapper[4823]: I0320 08:57:41.598252 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:41 crc kubenswrapper[4823]: I0320 08:57:41.598266 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:41 crc kubenswrapper[4823]: I0320 08:57:41.598910 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:41 crc kubenswrapper[4823]: E0320 08:57:41.606600 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:57:42 crc kubenswrapper[4823]: I0320 08:57:42.025764 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:43 crc kubenswrapper[4823]: I0320 08:57:43.024145 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:44 crc kubenswrapper[4823]: I0320 08:57:44.027638 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:44 crc kubenswrapper[4823]: E0320 08:57:44.153956 4823 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:57:45 crc kubenswrapper[4823]: I0320 08:57:45.027310 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:46 crc kubenswrapper[4823]: I0320 08:57:46.025348 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:46 crc kubenswrapper[4823]: I0320 08:57:46.762559 4823 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:59070->192.168.126.11:10357: read: connection reset by peer" start-of-body= Mar 20 08:57:46 crc kubenswrapper[4823]: I0320 08:57:46.762661 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:59070->192.168.126.11:10357: read: connection reset by peer" Mar 20 08:57:46 crc kubenswrapper[4823]: I0320 08:57:46.762788 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:46 crc kubenswrapper[4823]: I0320 08:57:46.763024 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:46 crc kubenswrapper[4823]: I0320 08:57:46.770204 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:46 crc kubenswrapper[4823]: I0320 08:57:46.770457 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:46 crc kubenswrapper[4823]: I0320 08:57:46.770621 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:46 crc kubenswrapper[4823]: I0320 08:57:46.771718 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 20 08:57:46 crc kubenswrapper[4823]: I0320 08:57:46.772277 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc" gracePeriod=30 Mar 20 08:57:46 crc kubenswrapper[4823]: E0320 08:57:46.780291 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 08:57:46 crc kubenswrapper[4823]: &Event{ObjectMeta:{kube-controller-manager-crc.189e80eee4f8c2a4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": read tcp 192.168.126.11:59070->192.168.126.11:10357: read: connection reset by peer Mar 20 08:57:46 crc kubenswrapper[4823]: body: Mar 20 08:57:46 crc kubenswrapper[4823]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:46.762633892 +0000 UTC m=+33.448530306,LastTimestamp:2026-03-20 08:57:46.762633892 +0000 UTC m=+33.448530306,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:46 crc kubenswrapper[4823]: > Mar 20 08:57:46 crc kubenswrapper[4823]: E0320 08:57:46.786063 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80eee4f9d5e2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:59070->192.168.126.11:10357: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:46.762704354 +0000 UTC m=+33.448600778,LastTimestamp:2026-03-20 08:57:46.762704354 +0000 UTC m=+33.448600778,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:46 crc kubenswrapper[4823]: E0320 08:57:46.793856 4823 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80eee58b536c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:46.772239212 +0000 UTC m=+33.458135656,LastTimestamp:2026-03-20 08:57:46.772239212 +0000 UTC m=+33.458135656,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:46 crc kubenswrapper[4823]: E0320 08:57:46.804376 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e78d9b01b2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e78d9b01b2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.232100786 +0000 UTC m=+1.917997160,LastTimestamp:2026-03-20 08:57:46.797785154 +0000 UTC m=+33.483681568,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:47 crc kubenswrapper[4823]: E0320 08:57:47.018427 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e79f9a2154\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e79f9a2154 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.534033236 +0000 UTC m=+2.219929610,LastTimestamp:2026-03-20 08:57:47.011904373 +0000 UTC m=+33.697800777,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:47 crc kubenswrapper[4823]: I0320 08:57:47.022714 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:47 crc kubenswrapper[4823]: E0320 08:57:47.029647 4823 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e7a074233e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e7a074233e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:15.548320574 +0000 UTC m=+2.234216948,LastTimestamp:2026-03-20 08:57:47.024082435 +0000 UTC m=+33.709978819,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:47 crc kubenswrapper[4823]: I0320 08:57:47.267396 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 08:57:47 crc kubenswrapper[4823]: I0320 08:57:47.268086 4823 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc" exitCode=255 Mar 20 08:57:47 crc kubenswrapper[4823]: I0320 08:57:47.268147 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc"} Mar 20 08:57:47 crc kubenswrapper[4823]: I0320 08:57:47.268187 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296"} Mar 20 08:57:47 crc kubenswrapper[4823]: I0320 08:57:47.268318 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:47 crc kubenswrapper[4823]: I0320 08:57:47.270551 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:47 crc kubenswrapper[4823]: I0320 08:57:47.270594 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:47 crc kubenswrapper[4823]: I0320 08:57:47.270612 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:48 crc kubenswrapper[4823]: I0320 08:57:48.023642 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:48 crc kubenswrapper[4823]: E0320 08:57:48.599799 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:57:48 crc kubenswrapper[4823]: I0320 08:57:48.607509 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:48 crc kubenswrapper[4823]: I0320 08:57:48.609346 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:48 crc kubenswrapper[4823]: I0320 08:57:48.609424 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:48 crc kubenswrapper[4823]: I0320 08:57:48.609454 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:48 crc kubenswrapper[4823]: I0320 08:57:48.609492 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:48 crc kubenswrapper[4823]: E0320 08:57:48.616547 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:57:49 crc kubenswrapper[4823]: I0320 08:57:49.026353 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:50 crc kubenswrapper[4823]: I0320 08:57:50.022708 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:51 crc kubenswrapper[4823]: I0320 08:57:51.025881 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:52 crc kubenswrapper[4823]: I0320 08:57:52.027226 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:52 crc kubenswrapper[4823]: I0320 08:57:52.097282 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:52 crc kubenswrapper[4823]: I0320 08:57:52.099344 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:52 crc kubenswrapper[4823]: I0320 08:57:52.099423 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:52 crc kubenswrapper[4823]: I0320 08:57:52.099444 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:52 crc kubenswrapper[4823]: I0320 08:57:52.100554 4823 scope.go:117] "RemoveContainer" containerID="3fda6b6e4da6f6cff2e64233e6e8daad78457d875abc4e1b379fcd8e672f642b" Mar 20 08:57:52 crc kubenswrapper[4823]: I0320 08:57:52.222444 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:52 crc kubenswrapper[4823]: I0320 08:57:52.222667 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:52 crc kubenswrapper[4823]: I0320 08:57:52.223952 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:52 crc kubenswrapper[4823]: I0320 08:57:52.224005 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:52 crc kubenswrapper[4823]: I0320 08:57:52.224022 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.023562 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:53 crc kubenswrapper[4823]: W0320 08:57:53.282502 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:53 crc kubenswrapper[4823]: E0320 08:57:53.282568 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.286731 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.287555 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.289846 4823 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="504e4645f3df1657634d3c4703b84dfa9be85900b2f76647b3e6d11ea2d02d8a" exitCode=255 Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.289900 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"504e4645f3df1657634d3c4703b84dfa9be85900b2f76647b3e6d11ea2d02d8a"} Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.289949 4823 scope.go:117] "RemoveContainer" containerID="3fda6b6e4da6f6cff2e64233e6e8daad78457d875abc4e1b379fcd8e672f642b" Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.290142 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.291168 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.291281 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.291346 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.292444 4823 scope.go:117] "RemoveContainer" containerID="504e4645f3df1657634d3c4703b84dfa9be85900b2f76647b3e6d11ea2d02d8a" Mar 20 08:57:53 crc kubenswrapper[4823]: E0320 08:57:53.292763 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:53 crc kubenswrapper[4823]: I0320 08:57:53.642683 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.024007 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:54 crc kubenswrapper[4823]: E0320 08:57:54.154162 4823 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.163870 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.164094 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.165188 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.165287 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.165366 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.168306 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.294372 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.296605 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.296617 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.297950 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.297994 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.298008 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.298916 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.298944 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.298955 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:54 crc kubenswrapper[4823]: I0320 08:57:54.299341 4823 scope.go:117] "RemoveContainer" containerID="504e4645f3df1657634d3c4703b84dfa9be85900b2f76647b3e6d11ea2d02d8a" Mar 20 08:57:54 crc kubenswrapper[4823]: E0320 08:57:54.299509 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:55 crc kubenswrapper[4823]: I0320 08:57:55.025098 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:55 crc kubenswrapper[4823]: E0320 08:57:55.605355 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:57:55 crc kubenswrapper[4823]: I0320 08:57:55.617476 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:55 crc kubenswrapper[4823]: I0320 08:57:55.618824 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:55 crc kubenswrapper[4823]: I0320 08:57:55.618874 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:55 crc kubenswrapper[4823]: I0320 08:57:55.618888 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:55 crc kubenswrapper[4823]: I0320 08:57:55.618918 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:55 crc kubenswrapper[4823]: E0320 08:57:55.633362 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:57:56 crc kubenswrapper[4823]: I0320 08:57:56.022978 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:57 crc kubenswrapper[4823]: I0320 08:57:57.024924 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:57 crc kubenswrapper[4823]: W0320 08:57:57.993158 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 20 08:57:57 crc kubenswrapper[4823]: E0320 08:57:57.993696 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 20 08:57:58 crc kubenswrapper[4823]: I0320 08:57:58.023722 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:59 crc kubenswrapper[4823]: I0320 08:57:59.024521 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:59 crc kubenswrapper[4823]: I0320 08:57:59.131411 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:59 crc kubenswrapper[4823]: I0320 08:57:59.131710 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:59 crc kubenswrapper[4823]: I0320 08:57:59.133593 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:59 crc kubenswrapper[4823]: I0320 08:57:59.133670 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:59 crc kubenswrapper[4823]: I0320 08:57:59.133691 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:59 crc kubenswrapper[4823]: I0320 08:57:59.134833 4823 scope.go:117] "RemoveContainer" containerID="504e4645f3df1657634d3c4703b84dfa9be85900b2f76647b3e6d11ea2d02d8a" Mar 20 08:57:59 crc kubenswrapper[4823]: E0320 08:57:59.135146 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:00 crc kubenswrapper[4823]: I0320 08:58:00.023309 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:01 crc kubenswrapper[4823]: I0320 08:58:01.024677 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:01 crc kubenswrapper[4823]: W0320 08:58:01.147691 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 20 08:58:01 crc kubenswrapper[4823]: E0320 08:58:01.147819 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 20 08:58:01 crc kubenswrapper[4823]: I0320 08:58:01.827190 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:58:01 crc kubenswrapper[4823]: I0320 08:58:01.827452 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:01 crc kubenswrapper[4823]: I0320 08:58:01.829295 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:01 crc kubenswrapper[4823]: I0320 08:58:01.829358 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:01 crc kubenswrapper[4823]: I0320 08:58:01.829381 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:02 crc kubenswrapper[4823]: I0320 08:58:02.026112 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:02 crc kubenswrapper[4823]: I0320 08:58:02.227997 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:58:02 crc kubenswrapper[4823]: I0320 08:58:02.228228 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:02 crc kubenswrapper[4823]: I0320 08:58:02.229957 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:02 crc kubenswrapper[4823]: I0320 08:58:02.230053 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:02 crc kubenswrapper[4823]: I0320 08:58:02.230073 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:02 crc kubenswrapper[4823]: E0320 08:58:02.609609 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:58:02 crc kubenswrapper[4823]: I0320 08:58:02.634139 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:02 crc kubenswrapper[4823]: I0320 08:58:02.635949 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:02 crc kubenswrapper[4823]: I0320 08:58:02.636026 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:02 crc kubenswrapper[4823]: I0320 08:58:02.636040 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:02 crc kubenswrapper[4823]: I0320 08:58:02.636081 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:58:02 crc kubenswrapper[4823]: E0320 08:58:02.642059 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:58:03 crc kubenswrapper[4823]: I0320 08:58:03.026244 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:04 crc kubenswrapper[4823]: I0320 08:58:04.025583 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:04 crc kubenswrapper[4823]: E0320 08:58:04.154281 4823 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:58:05 crc kubenswrapper[4823]: I0320 08:58:05.022895 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:06 crc kubenswrapper[4823]: I0320 08:58:06.024635 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:06 crc kubenswrapper[4823]: W0320 08:58:06.270882 4823 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 20 08:58:06 crc kubenswrapper[4823]: E0320 08:58:06.271009 4823 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 20 08:58:07 crc kubenswrapper[4823]: I0320 08:58:07.025926 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:08 crc kubenswrapper[4823]: I0320 08:58:08.025717 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:09 crc kubenswrapper[4823]: I0320 08:58:09.025642 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:09 crc kubenswrapper[4823]: E0320 08:58:09.617967 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:58:09 crc kubenswrapper[4823]: I0320 08:58:09.642291 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:09 crc kubenswrapper[4823]: I0320 08:58:09.644472 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:09 crc kubenswrapper[4823]: I0320 08:58:09.644772 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:09 crc kubenswrapper[4823]: I0320 08:58:09.644960 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:09 crc kubenswrapper[4823]: I0320 08:58:09.645179 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:58:09 crc kubenswrapper[4823]: E0320 08:58:09.654338 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:58:10 crc kubenswrapper[4823]: I0320 08:58:10.026956 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:11 crc kubenswrapper[4823]: I0320 08:58:11.025069 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:12 crc kubenswrapper[4823]: I0320 08:58:12.025916 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:13 crc kubenswrapper[4823]: I0320 08:58:13.025644 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.023544 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.097098 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.098549 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.098587 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.098598 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.099124 4823 scope.go:117] "RemoveContainer" containerID="504e4645f3df1657634d3c4703b84dfa9be85900b2f76647b3e6d11ea2d02d8a" Mar 20 08:58:14 crc kubenswrapper[4823]: E0320 08:58:14.154494 4823 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.356328 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.358450 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099"} Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.358608 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.359932 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.359974 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:14 crc kubenswrapper[4823]: I0320 08:58:14.359990 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:15 crc kubenswrapper[4823]: I0320 08:58:15.022016 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:15 crc kubenswrapper[4823]: I0320 08:58:15.363566 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 08:58:15 crc kubenswrapper[4823]: I0320 08:58:15.365053 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 08:58:15 crc kubenswrapper[4823]: I0320 08:58:15.367225 4823 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099" exitCode=255 Mar 20 08:58:15 crc kubenswrapper[4823]: I0320 08:58:15.367296 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099"} Mar 20 08:58:15 crc kubenswrapper[4823]: I0320 08:58:15.367349 4823 scope.go:117] "RemoveContainer" containerID="504e4645f3df1657634d3c4703b84dfa9be85900b2f76647b3e6d11ea2d02d8a" Mar 20 08:58:15 crc kubenswrapper[4823]: I0320 08:58:15.367602 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:15 crc kubenswrapper[4823]: I0320 08:58:15.369380 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:15 crc kubenswrapper[4823]: I0320 08:58:15.369418 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:15 crc kubenswrapper[4823]: I0320 08:58:15.369431 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:15 crc kubenswrapper[4823]: I0320 08:58:15.370072 4823 scope.go:117] "RemoveContainer" containerID="6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099" Mar 20 08:58:15 crc kubenswrapper[4823]: E0320 08:58:15.370262 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:16 crc kubenswrapper[4823]: I0320 08:58:16.026921 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:16 crc kubenswrapper[4823]: I0320 08:58:16.372128 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 08:58:16 crc kubenswrapper[4823]: E0320 08:58:16.627227 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:58:16 crc kubenswrapper[4823]: I0320 08:58:16.655264 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:16 crc kubenswrapper[4823]: I0320 08:58:16.657511 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:16 crc kubenswrapper[4823]: I0320 08:58:16.657802 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:16 crc kubenswrapper[4823]: I0320 08:58:16.657824 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:16 crc kubenswrapper[4823]: I0320 08:58:16.657866 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:58:16 crc kubenswrapper[4823]: E0320 08:58:16.664790 4823 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:58:17 crc kubenswrapper[4823]: I0320 08:58:17.025148 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:18 crc kubenswrapper[4823]: I0320 08:58:18.024890 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:19 crc kubenswrapper[4823]: I0320 08:58:19.022524 4823 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:19 crc kubenswrapper[4823]: I0320 08:58:19.118362 4823 csr.go:261] certificate signing request csr-9gjpl is approved, waiting to be issued Mar 20 08:58:19 crc kubenswrapper[4823]: I0320 08:58:19.127134 4823 csr.go:257] certificate signing request csr-9gjpl is issued Mar 20 08:58:19 crc kubenswrapper[4823]: I0320 08:58:19.131175 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:58:19 crc kubenswrapper[4823]: I0320 08:58:19.131451 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:19 crc kubenswrapper[4823]: I0320 08:58:19.132874 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:19 crc kubenswrapper[4823]: I0320 08:58:19.132995 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:19 crc kubenswrapper[4823]: I0320 08:58:19.133068 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:19 crc kubenswrapper[4823]: I0320 08:58:19.133732 4823 scope.go:117] "RemoveContainer" containerID="6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099" Mar 20 08:58:19 crc kubenswrapper[4823]: E0320 08:58:19.134012 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:19 crc kubenswrapper[4823]: I0320 08:58:19.219323 4823 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 20 08:58:19 crc kubenswrapper[4823]: I0320 08:58:19.871909 4823 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 20 08:58:20 crc kubenswrapper[4823]: I0320 08:58:20.129360 4823 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-08 16:06:20.870753576 +0000 UTC Mar 20 08:58:20 crc kubenswrapper[4823]: I0320 08:58:20.129411 4823 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6319h8m0.741345461s for next certificate rotation Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.642637 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.642839 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.644148 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.644200 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.644211 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.644934 4823 scope.go:117] "RemoveContainer" containerID="6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099" Mar 20 08:58:23 crc kubenswrapper[4823]: E0320 08:58:23.645105 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.664971 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.666574 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.666636 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.666662 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.667068 4823 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.676286 4823 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.676682 4823 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 20 08:58:23 crc kubenswrapper[4823]: E0320 08:58:23.676722 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.683611 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.683692 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.683706 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.683730 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.683790 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:23Z","lastTransitionTime":"2026-03-20T08:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:23 crc kubenswrapper[4823]: E0320 08:58:23.704420 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.715247 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.715337 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.715366 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.715400 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.715425 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:23Z","lastTransitionTime":"2026-03-20T08:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:23 crc kubenswrapper[4823]: E0320 08:58:23.732470 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.741261 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.741295 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.741306 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.741321 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.741331 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:23Z","lastTransitionTime":"2026-03-20T08:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:23 crc kubenswrapper[4823]: E0320 08:58:23.759365 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.771830 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.771870 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.771879 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.771894 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:23 crc kubenswrapper[4823]: I0320 08:58:23.771907 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:23Z","lastTransitionTime":"2026-03-20T08:58:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:23 crc kubenswrapper[4823]: E0320 08:58:23.791133 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4823]: E0320 08:58:23.791307 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:58:23 crc kubenswrapper[4823]: E0320 08:58:23.791339 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:23 crc kubenswrapper[4823]: E0320 08:58:23.892366 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:23 crc kubenswrapper[4823]: E0320 08:58:23.993332 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:24 crc kubenswrapper[4823]: E0320 08:58:24.093720 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:24 crc kubenswrapper[4823]: E0320 08:58:24.156290 4823 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:58:24 crc kubenswrapper[4823]: E0320 08:58:24.194386 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:24 crc kubenswrapper[4823]: E0320 08:58:24.295233 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:24 crc kubenswrapper[4823]: E0320 08:58:24.395889 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:24 crc kubenswrapper[4823]: E0320 08:58:24.496290 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:24 crc kubenswrapper[4823]: E0320 08:58:24.597221 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:24 crc kubenswrapper[4823]: E0320 08:58:24.697734 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:24 crc kubenswrapper[4823]: E0320 08:58:24.798137 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:24 crc kubenswrapper[4823]: E0320 08:58:24.898415 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:24 crc kubenswrapper[4823]: E0320 08:58:24.999400 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:25 crc kubenswrapper[4823]: E0320 08:58:25.100299 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:25 crc kubenswrapper[4823]: E0320 08:58:25.201102 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:25 crc kubenswrapper[4823]: E0320 08:58:25.301534 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:25 crc kubenswrapper[4823]: E0320 08:58:25.401804 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:25 crc kubenswrapper[4823]: E0320 08:58:25.503293 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:25 crc kubenswrapper[4823]: E0320 08:58:25.604484 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:25 crc kubenswrapper[4823]: E0320 08:58:25.705535 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:25 crc kubenswrapper[4823]: E0320 08:58:25.806686 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:25 crc kubenswrapper[4823]: E0320 08:58:25.907313 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:26 crc kubenswrapper[4823]: E0320 08:58:26.007916 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:26 crc kubenswrapper[4823]: E0320 08:58:26.108607 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:26 crc kubenswrapper[4823]: E0320 08:58:26.209449 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:26 crc kubenswrapper[4823]: E0320 08:58:26.309805 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:26 crc kubenswrapper[4823]: E0320 08:58:26.410798 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:26 crc kubenswrapper[4823]: E0320 08:58:26.511502 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:26 crc kubenswrapper[4823]: E0320 08:58:26.612256 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:26 crc kubenswrapper[4823]: E0320 08:58:26.712608 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:26 crc kubenswrapper[4823]: E0320 08:58:26.813631 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:26 crc kubenswrapper[4823]: E0320 08:58:26.914556 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:27 crc kubenswrapper[4823]: E0320 08:58:27.015584 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:27 crc kubenswrapper[4823]: E0320 08:58:27.116225 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:27 crc kubenswrapper[4823]: E0320 08:58:27.216899 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:27 crc kubenswrapper[4823]: E0320 08:58:27.317302 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:27 crc kubenswrapper[4823]: E0320 08:58:27.417812 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:27 crc kubenswrapper[4823]: E0320 08:58:27.518698 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:27 crc kubenswrapper[4823]: E0320 08:58:27.619002 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:27 crc kubenswrapper[4823]: E0320 08:58:27.719763 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:27 crc kubenswrapper[4823]: E0320 08:58:27.820514 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:27 crc kubenswrapper[4823]: E0320 08:58:27.921636 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:28 crc kubenswrapper[4823]: E0320 08:58:28.022084 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:28 crc kubenswrapper[4823]: E0320 08:58:28.123043 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:28 crc kubenswrapper[4823]: E0320 08:58:28.223280 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:28 crc kubenswrapper[4823]: I0320 08:58:28.223473 4823 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 20 08:58:28 crc kubenswrapper[4823]: E0320 08:58:28.324840 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:28 crc kubenswrapper[4823]: E0320 08:58:28.425466 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:28 crc kubenswrapper[4823]: E0320 08:58:28.526049 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:28 crc kubenswrapper[4823]: E0320 08:58:28.627366 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:28 crc kubenswrapper[4823]: E0320 08:58:28.728706 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:28 crc kubenswrapper[4823]: E0320 08:58:28.829210 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:28 crc kubenswrapper[4823]: E0320 08:58:28.929836 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:29 crc kubenswrapper[4823]: E0320 08:58:29.029978 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:29 crc kubenswrapper[4823]: E0320 08:58:29.130583 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:29 crc kubenswrapper[4823]: E0320 08:58:29.230991 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:29 crc kubenswrapper[4823]: E0320 08:58:29.331465 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:29 crc kubenswrapper[4823]: E0320 08:58:29.432185 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:29 crc kubenswrapper[4823]: E0320 08:58:29.533269 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:29 crc kubenswrapper[4823]: E0320 08:58:29.634335 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:29 crc kubenswrapper[4823]: E0320 08:58:29.735433 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:29 crc kubenswrapper[4823]: E0320 08:58:29.836288 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:29 crc kubenswrapper[4823]: E0320 08:58:29.937817 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:30 crc kubenswrapper[4823]: E0320 08:58:30.038421 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:30 crc kubenswrapper[4823]: E0320 08:58:30.139518 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:30 crc kubenswrapper[4823]: E0320 08:58:30.240214 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:30 crc kubenswrapper[4823]: E0320 08:58:30.340610 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:30 crc kubenswrapper[4823]: E0320 08:58:30.441633 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:30 crc kubenswrapper[4823]: E0320 08:58:30.542352 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:30 crc kubenswrapper[4823]: E0320 08:58:30.643826 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:30 crc kubenswrapper[4823]: E0320 08:58:30.744055 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:30 crc kubenswrapper[4823]: E0320 08:58:30.844815 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:30 crc kubenswrapper[4823]: E0320 08:58:30.945291 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:31 crc kubenswrapper[4823]: E0320 08:58:31.045473 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:31 crc kubenswrapper[4823]: E0320 08:58:31.146183 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:31 crc kubenswrapper[4823]: E0320 08:58:31.246662 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:31 crc kubenswrapper[4823]: E0320 08:58:31.347237 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:31 crc kubenswrapper[4823]: E0320 08:58:31.447774 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:31 crc kubenswrapper[4823]: E0320 08:58:31.548163 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:31 crc kubenswrapper[4823]: E0320 08:58:31.649294 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:31 crc kubenswrapper[4823]: E0320 08:58:31.750366 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:31 crc kubenswrapper[4823]: E0320 08:58:31.851936 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:31 crc kubenswrapper[4823]: E0320 08:58:31.953227 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:32 crc kubenswrapper[4823]: E0320 08:58:32.053359 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:32 crc kubenswrapper[4823]: E0320 08:58:32.154538 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:32 crc kubenswrapper[4823]: E0320 08:58:32.255252 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:32 crc kubenswrapper[4823]: E0320 08:58:32.355976 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:32 crc kubenswrapper[4823]: E0320 08:58:32.456393 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:32 crc kubenswrapper[4823]: E0320 08:58:32.557855 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:32 crc kubenswrapper[4823]: E0320 08:58:32.658820 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:32 crc kubenswrapper[4823]: E0320 08:58:32.759049 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:32 crc kubenswrapper[4823]: E0320 08:58:32.859198 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:32 crc kubenswrapper[4823]: E0320 08:58:32.959969 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.060221 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.161524 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.262282 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.362477 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.463429 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.564584 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.664858 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.765323 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.816912 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.823408 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.823462 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.823476 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.823494 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.823506 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.840814 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.846166 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.846232 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.846245 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.846276 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.846291 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.866205 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.871474 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.871543 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.871560 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.871587 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.871607 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.887792 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.892007 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.892061 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.892077 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.892096 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4823]: I0320 08:58:33.892109 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.904275 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.904427 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:58:33 crc kubenswrapper[4823]: E0320 08:58:33.904460 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:34 crc kubenswrapper[4823]: E0320 08:58:34.004554 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:34 crc kubenswrapper[4823]: E0320 08:58:34.105213 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:34 crc kubenswrapper[4823]: E0320 08:58:34.157708 4823 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:58:34 crc kubenswrapper[4823]: E0320 08:58:34.205341 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:34 crc kubenswrapper[4823]: E0320 08:58:34.306517 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:34 crc kubenswrapper[4823]: E0320 08:58:34.407677 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:34 crc kubenswrapper[4823]: E0320 08:58:34.508854 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:34 crc kubenswrapper[4823]: E0320 08:58:34.609642 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:34 crc kubenswrapper[4823]: E0320 08:58:34.710665 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:34 crc kubenswrapper[4823]: E0320 08:58:34.811102 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:34 crc kubenswrapper[4823]: E0320 08:58:34.911961 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:35 crc kubenswrapper[4823]: E0320 08:58:35.012856 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:35 crc kubenswrapper[4823]: I0320 08:58:35.096343 4823 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:35 crc kubenswrapper[4823]: I0320 08:58:35.097653 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:35 crc kubenswrapper[4823]: I0320 08:58:35.097690 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:35 crc kubenswrapper[4823]: I0320 08:58:35.097700 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:35 crc kubenswrapper[4823]: I0320 08:58:35.098347 4823 scope.go:117] "RemoveContainer" containerID="6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099" Mar 20 08:58:35 crc kubenswrapper[4823]: E0320 08:58:35.098532 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:35 crc kubenswrapper[4823]: E0320 08:58:35.113007 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:35 crc kubenswrapper[4823]: E0320 08:58:35.213589 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:35 crc kubenswrapper[4823]: E0320 08:58:35.314242 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:35 crc kubenswrapper[4823]: E0320 08:58:35.415283 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:35 crc kubenswrapper[4823]: E0320 08:58:35.516246 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:35 crc kubenswrapper[4823]: E0320 08:58:35.617072 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:35 crc kubenswrapper[4823]: E0320 08:58:35.718127 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:35 crc kubenswrapper[4823]: E0320 08:58:35.818267 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:35 crc kubenswrapper[4823]: E0320 08:58:35.918633 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:36 crc kubenswrapper[4823]: E0320 08:58:36.019309 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:36 crc kubenswrapper[4823]: E0320 08:58:36.119817 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:36 crc kubenswrapper[4823]: E0320 08:58:36.220345 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:36 crc kubenswrapper[4823]: E0320 08:58:36.320492 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:36 crc kubenswrapper[4823]: E0320 08:58:36.421091 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:36 crc kubenswrapper[4823]: E0320 08:58:36.521698 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:36 crc kubenswrapper[4823]: E0320 08:58:36.622918 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:36 crc kubenswrapper[4823]: E0320 08:58:36.723361 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:36 crc kubenswrapper[4823]: E0320 08:58:36.823911 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:36 crc kubenswrapper[4823]: E0320 08:58:36.924088 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:37 crc kubenswrapper[4823]: E0320 08:58:37.024797 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:37 crc kubenswrapper[4823]: E0320 08:58:37.125241 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:37 crc kubenswrapper[4823]: E0320 08:58:37.225808 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:37 crc kubenswrapper[4823]: E0320 08:58:37.327009 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:37 crc kubenswrapper[4823]: E0320 08:58:37.428193 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:37 crc kubenswrapper[4823]: E0320 08:58:37.529120 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:37 crc kubenswrapper[4823]: E0320 08:58:37.630171 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:37 crc kubenswrapper[4823]: I0320 08:58:37.685597 4823 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 20 08:58:37 crc kubenswrapper[4823]: E0320 08:58:37.731195 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:37 crc kubenswrapper[4823]: E0320 08:58:37.831359 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:37 crc kubenswrapper[4823]: E0320 08:58:37.931500 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:38 crc kubenswrapper[4823]: E0320 08:58:38.032038 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:38 crc kubenswrapper[4823]: E0320 08:58:38.133186 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:38 crc kubenswrapper[4823]: E0320 08:58:38.233672 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:38 crc kubenswrapper[4823]: E0320 08:58:38.334347 4823 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.378977 4823 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.439161 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.439236 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.439256 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.439284 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.439307 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.544020 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.544091 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.544111 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.544140 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.544161 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.648077 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.648174 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.648199 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.648231 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.648258 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.751644 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.751733 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.751781 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.751806 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.751823 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.855306 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.855369 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.855385 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.855406 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.855422 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.959180 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.959244 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.959259 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.959280 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4823]: I0320 08:58:38.959295 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.045538 4823 apiserver.go:52] "Watching apiserver" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.053349 4823 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.053697 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-7wn7x","openshift-machine-config-operator/machine-config-daemon-snzb2","openshift-multus/network-metrics-daemon-jzzk2","openshift-network-operator/iptables-alerter-4ln5h","openshift-multus/multus-9vsj7","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq","openshift-ovn-kubernetes/ovnkube-node-2x54k","openshift-dns/node-resolver-z4b65","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-multus/multus-additional-cni-plugins-fb4mg","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.054151 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.054151 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.054210 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.054291 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.054347 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.054458 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.054538 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.054514 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.054938 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7wn7x" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.054963 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.056344 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.058793 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.059794 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.059969 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.062599 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.062783 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.062825 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.063024 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.063520 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.063591 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-z4b65" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.063813 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.063992 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.064178 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.064260 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.064298 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.064305 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.064448 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.064681 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.064711 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.064720 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.064736 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.064762 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.064783 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.066767 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.066869 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.066891 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.067022 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.067099 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.067341 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.067934 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.068033 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.068093 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.068277 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.068359 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.068497 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.068649 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.070677 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.070906 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.071047 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.071301 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.071394 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.071480 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.071515 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.071340 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.071111 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.071674 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.071729 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.071949 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.072115 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.072339 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.084523 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.094480 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.103581 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.116699 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.123958 4823 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.128138 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.136509 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145428 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145480 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145511 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145539 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145564 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145588 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145610 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145632 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145656 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145676 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145695 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145715 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145732 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145765 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145789 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145815 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145836 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145857 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145853 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145879 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145900 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145920 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145941 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145962 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.145982 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146002 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146021 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146043 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146064 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146086 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146106 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146127 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146148 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146166 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146169 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146185 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146204 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146220 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146241 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146259 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146278 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146297 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146315 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146334 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146351 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146369 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146387 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146407 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146428 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146427 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146450 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146469 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146486 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146503 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146520 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146536 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146552 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146567 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146585 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146602 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146620 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146638 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146648 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146656 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146674 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146694 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146712 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146731 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146768 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146789 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146814 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146835 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.146881 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147080 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147600 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147650 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147667 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147708 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147729 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147805 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147867 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147890 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147899 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147908 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147954 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.147980 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148009 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148033 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148054 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148068 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148097 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148195 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148246 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148454 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148276 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148547 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148587 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148612 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148640 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148665 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148692 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148721 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148776 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148805 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148805 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148837 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148872 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148906 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148932 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148956 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.148991 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149022 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149078 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149085 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149083 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149093 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149087 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149248 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149278 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149314 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149341 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149383 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149410 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149446 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149454 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149494 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149533 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149628 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149921 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149913 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.150258 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.150397 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.150480 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.150495 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.150579 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.150968 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.151103 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.149481 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.152543 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.151498 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.151662 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.151697 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153020 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.152004 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153067 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153123 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153157 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153194 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153227 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153256 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153285 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153315 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153340 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153364 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153398 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153428 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153455 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153484 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153522 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153556 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153658 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153690 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153718 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153773 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153802 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153832 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153859 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153893 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153936 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153972 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154006 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154038 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154084 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154122 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.152154 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154168 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154211 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154248 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154285 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154321 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154356 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154393 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154431 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154470 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154509 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154548 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154585 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154621 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154659 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154696 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154728 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154816 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154856 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154889 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154913 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154937 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154963 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155004 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155048 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155084 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155118 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155157 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155195 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155233 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155281 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155306 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155335 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155360 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155386 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155414 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155441 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155469 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155573 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155604 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155632 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155658 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155690 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155715 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155804 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155832 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155862 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155889 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155913 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155937 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155960 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155992 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156029 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156064 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156104 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156196 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156232 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-cni-dir\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156257 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-systemd\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156287 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156327 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156355 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156384 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156418 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156452 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-cnibin\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156484 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gfvd\" (UniqueName: \"kubernetes.io/projected/38298bb6-9157-4c18-94a2-fedec7e95a4c-kube-api-access-2gfvd\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156513 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dllqz\" (UniqueName: \"kubernetes.io/projected/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-kube-api-access-dllqz\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156542 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156571 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-hostroot\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156596 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-system-cni-dir\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156625 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-os-release\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156649 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovn-node-metrics-cert\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156674 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwkjs\" (UniqueName: \"kubernetes.io/projected/1bed1862-5bcd-46fe-bc84-2dbab416086b-kube-api-access-mwkjs\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156698 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-config\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156728 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvgxr\" (UniqueName: \"kubernetes.io/projected/7748c732-655d-42c5-9dee-f7c69315471b-kube-api-access-tvgxr\") pod \"node-resolver-z4b65\" (UID: \"7748c732-655d-42c5-9dee-f7c69315471b\") " pod="openshift-dns/node-resolver-z4b65" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156781 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-run-netns\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156804 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-systemd-units\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156828 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-ovn-kubernetes\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156853 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38298bb6-9157-4c18-94a2-fedec7e95a4c-cni-binary-copy\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156886 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156917 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6c32331c-bb6a-475a-847f-40b25351917e-host\") pod \"node-ca-7wn7x\" (UID: \"6c32331c-bb6a-475a-847f-40b25351917e\") " pod="openshift-image-registry/node-ca-7wn7x" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156941 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-cnibin\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156972 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157004 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7322ac31-c6c4-43b9-8b3c-60ff84742545-rootfs\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157062 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-var-lib-cni-bin\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157098 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4wm2\" (UniqueName: \"kubernetes.io/projected/316f49ed-2e3c-42b7-a133-a36c3074810e-kube-api-access-p4wm2\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157135 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38298bb6-9157-4c18-94a2-fedec7e95a4c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157170 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157194 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-ovn\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157218 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-log-socket\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157244 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-netd\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157267 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-etc-kubernetes\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157293 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zf2d\" (UniqueName: \"kubernetes.io/projected/d33f9d63-34d7-4722-95d5-13bdcb32b99a-kube-api-access-2zf2d\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157317 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-slash\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157338 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-etc-openvswitch\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157364 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7322ac31-c6c4-43b9-8b3c-60ff84742545-proxy-tls\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157395 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157421 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-system-cni-dir\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157447 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-os-release\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157476 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-conf-dir\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157509 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-netns\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157542 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157571 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6c32331c-bb6a-475a-847f-40b25351917e-serviceca\") pod \"node-ca-7wn7x\" (UID: \"6c32331c-bb6a-475a-847f-40b25351917e\") " pod="openshift-image-registry/node-ca-7wn7x" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157600 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-openvswitch\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157630 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-node-log\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157893 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157923 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7748c732-655d-42c5-9dee-f7c69315471b-hosts-file\") pod \"node-resolver-z4b65\" (UID: \"7748c732-655d-42c5-9dee-f7c69315471b\") " pod="openshift-dns/node-resolver-z4b65" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157956 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157985 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158023 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158057 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-var-lib-openvswitch\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158093 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-run-multus-certs\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158124 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158158 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7322ac31-c6c4-43b9-8b3c-60ff84742545-mcd-auth-proxy-config\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158190 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-socket-dir-parent\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158219 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-run-k8s-cni-cncf-io\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158253 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-daemon-config\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158282 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-var-lib-cni-multus\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158316 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-var-lib-kubelet\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158344 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-bin\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158370 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-env-overrides\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158408 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158438 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d33f9d63-34d7-4722-95d5-13bdcb32b99a-cni-binary-copy\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158467 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158499 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-kubelet\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158527 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-script-lib\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158558 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgm2f\" (UniqueName: \"kubernetes.io/projected/7322ac31-c6c4-43b9-8b3c-60ff84742545-kube-api-access-cgm2f\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158586 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgn6c\" (UniqueName: \"kubernetes.io/projected/6c32331c-bb6a-475a-847f-40b25351917e-kube-api-access-hgn6c\") pod \"node-ca-7wn7x\" (UID: \"6c32331c-bb6a-475a-847f-40b25351917e\") " pod="openshift-image-registry/node-ca-7wn7x" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158624 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158729 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158790 4823 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158812 4823 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158834 4823 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158856 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158875 4823 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158895 4823 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158915 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158933 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158951 4823 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158969 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158986 4823 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159004 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159027 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159044 4823 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159062 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159081 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159100 4823 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159118 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159138 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159161 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159178 4823 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159197 4823 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159216 4823 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159237 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159257 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159275 4823 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159293 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159312 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159331 4823 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159350 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159368 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159504 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159525 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.160915 4823 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.152177 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.152205 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.152256 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.152533 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.152623 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.152851 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153057 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153628 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153648 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.163629 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153687 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.153700 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154022 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154022 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154042 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154244 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154507 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154530 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.154672 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155184 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155403 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155804 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155842 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.155900 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156071 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156160 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156214 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.156924 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157168 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157523 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157543 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.157961 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158013 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.158497 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159143 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.164204 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159160 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159243 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.159785 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.164268 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.161913 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.162226 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.162584 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.164374 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.162849 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.162938 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.162990 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.163209 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.163226 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.163425 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.163559 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.164216 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.164721 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.164717 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.165157 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.165186 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.165264 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.152373 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.165602 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.165863 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.165730 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.166024 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.166122 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.167251 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.167560 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.166038 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.166709 4823 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.167725 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.167869 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:39.667803955 +0000 UTC m=+86.353700479 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.165388 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.168171 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.168258 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.168835 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.169608 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.169645 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.170246 4823 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.184212 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.184277 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:39.684229222 +0000 UTC m=+86.370125596 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.184309 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.184483 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.184521 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.184500 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.184659 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.179044 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.179170 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.179395 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.179417 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.184848 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.183704 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.183901 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.183908 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.183951 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.184972 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.170671 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.184648 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.185040 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.185297 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.185334 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.185353 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.185461 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.185451 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.185661 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.185681 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.186035 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.186035 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.186182 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.186188 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:58:39.686154463 +0000 UTC m=+86.372050837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.183846 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187178 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187194 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187216 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187232 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187247 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187264 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187303 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187328 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187432 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.188873 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187735 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187430 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.186701 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.186888 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.186886 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.186953 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187013 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187004 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187604 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187610 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.186564 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187702 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187810 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.187811 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.188431 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.188484 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.188680 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.188993 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.189003 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.189207 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.189226 4823 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.189307 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:39.689285834 +0000 UTC m=+86.375182208 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.189307 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.189623 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.188980 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.188702 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.189990 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.190012 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.191132 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.192106 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.192397 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.194592 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.194833 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.194871 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.194913 4823 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.195063 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:39.695020467 +0000 UTC m=+86.380916851 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.195835 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.195859 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.196102 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.196139 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.198028 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.198069 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.198732 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.198830 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.201200 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.201448 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.201475 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.201614 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.201817 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.202903 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.202996 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.203139 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.203325 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.203695 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.203893 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.204485 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.204626 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.204648 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.204920 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.205349 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.205580 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.205717 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.206109 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.206196 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.207294 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.207531 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.207773 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.216240 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.216235 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.226488 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.227434 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.231237 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.236496 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.269077 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272362 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-ovn-kubernetes\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272433 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38298bb6-9157-4c18-94a2-fedec7e95a4c-cni-binary-copy\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272471 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6c32331c-bb6a-475a-847f-40b25351917e-host\") pod \"node-ca-7wn7x\" (UID: \"6c32331c-bb6a-475a-847f-40b25351917e\") " pod="openshift-image-registry/node-ca-7wn7x" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272503 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-cnibin\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272532 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-systemd-units\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272555 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7322ac31-c6c4-43b9-8b3c-60ff84742545-rootfs\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272576 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-var-lib-cni-bin\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272602 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38298bb6-9157-4c18-94a2-fedec7e95a4c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272636 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-ovn\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272660 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-log-socket\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272683 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-netd\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272711 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4wm2\" (UniqueName: \"kubernetes.io/projected/316f49ed-2e3c-42b7-a133-a36c3074810e-kube-api-access-p4wm2\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272760 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zf2d\" (UniqueName: \"kubernetes.io/projected/d33f9d63-34d7-4722-95d5-13bdcb32b99a-kube-api-access-2zf2d\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272789 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-slash\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272813 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-etc-openvswitch\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272835 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7322ac31-c6c4-43b9-8b3c-60ff84742545-proxy-tls\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272859 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-system-cni-dir\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272884 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-etc-kubernetes\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272911 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-os-release\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272936 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-conf-dir\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272959 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-netns\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.272984 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6c32331c-bb6a-475a-847f-40b25351917e-serviceca\") pod \"node-ca-7wn7x\" (UID: \"6c32331c-bb6a-475a-847f-40b25351917e\") " pod="openshift-image-registry/node-ca-7wn7x" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273008 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-openvswitch\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273031 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-node-log\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273056 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7748c732-655d-42c5-9dee-f7c69315471b-hosts-file\") pod \"node-resolver-z4b65\" (UID: \"7748c732-655d-42c5-9dee-f7c69315471b\") " pod="openshift-dns/node-resolver-z4b65" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273080 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273107 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273142 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-var-lib-openvswitch\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273190 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273215 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7322ac31-c6c4-43b9-8b3c-60ff84742545-mcd-auth-proxy-config\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273240 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-socket-dir-parent\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273266 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-run-k8s-cni-cncf-io\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273289 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-daemon-config\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273311 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-run-multus-certs\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273342 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-var-lib-cni-multus\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273375 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-var-lib-kubelet\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273391 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-bin\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273412 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-env-overrides\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273430 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d33f9d63-34d7-4722-95d5-13bdcb32b99a-cni-binary-copy\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273449 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273464 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-script-lib\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273482 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgm2f\" (UniqueName: \"kubernetes.io/projected/7322ac31-c6c4-43b9-8b3c-60ff84742545-kube-api-access-cgm2f\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273501 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgn6c\" (UniqueName: \"kubernetes.io/projected/6c32331c-bb6a-475a-847f-40b25351917e-kube-api-access-hgn6c\") pod \"node-ca-7wn7x\" (UID: \"6c32331c-bb6a-475a-847f-40b25351917e\") " pod="openshift-image-registry/node-ca-7wn7x" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273517 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-kubelet\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273552 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-cni-dir\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273567 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-systemd\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273582 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273605 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273621 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273636 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-cnibin\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273655 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dllqz\" (UniqueName: \"kubernetes.io/projected/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-kube-api-access-dllqz\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273670 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273691 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-hostroot\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273720 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-system-cni-dir\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273765 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gfvd\" (UniqueName: \"kubernetes.io/projected/38298bb6-9157-4c18-94a2-fedec7e95a4c-kube-api-access-2gfvd\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273786 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-os-release\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273802 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovn-node-metrics-cert\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273823 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwkjs\" (UniqueName: \"kubernetes.io/projected/1bed1862-5bcd-46fe-bc84-2dbab416086b-kube-api-access-mwkjs\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273847 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvgxr\" (UniqueName: \"kubernetes.io/projected/7748c732-655d-42c5-9dee-f7c69315471b-kube-api-access-tvgxr\") pod \"node-resolver-z4b65\" (UID: \"7748c732-655d-42c5-9dee-f7c69315471b\") " pod="openshift-dns/node-resolver-z4b65" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273863 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-run-netns\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273879 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-config\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273945 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273961 4823 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273971 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273981 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.273992 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274003 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274013 4823 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274023 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274034 4823 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274046 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274059 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274071 4823 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274080 4823 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274091 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274099 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274110 4823 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274122 4823 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274131 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274140 4823 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274150 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274160 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274171 4823 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274182 4823 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274194 4823 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274204 4823 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274215 4823 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274228 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274240 4823 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274250 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274259 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274270 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274283 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274294 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274304 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274315 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274325 4823 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274335 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274355 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274365 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274374 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274385 4823 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274395 4823 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274404 4823 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274418 4823 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274429 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274441 4823 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274455 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274465 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274476 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274486 4823 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274496 4823 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274507 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274515 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274525 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274534 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274542 4823 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274554 4823 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274566 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274578 4823 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274588 4823 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274596 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274606 4823 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274616 4823 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274626 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274635 4823 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274644 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274652 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274660 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274669 4823 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274678 4823 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274687 4823 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274696 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274705 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274714 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274723 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274732 4823 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274760 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274771 4823 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274781 4823 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274790 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274799 4823 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274807 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274821 4823 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274829 4823 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274838 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274848 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274857 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274866 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274874 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274885 4823 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274896 4823 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274908 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274920 4823 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274931 4823 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274940 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274949 4823 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274958 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274967 4823 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274975 4823 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274984 4823 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.274993 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275002 4823 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275010 4823 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275020 4823 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275029 4823 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275038 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275048 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275056 4823 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275065 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275074 4823 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275083 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275091 4823 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275099 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275109 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275120 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275130 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275140 4823 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275149 4823 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275158 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275168 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275177 4823 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275186 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275195 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275206 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275214 4823 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275223 4823 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275232 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275241 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275249 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275258 4823 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275267 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275275 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275285 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275293 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275302 4823 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275312 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275322 4823 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275330 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275340 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275348 4823 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275357 4823 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275365 4823 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275374 4823 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275383 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275391 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275400 4823 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275412 4823 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275424 4823 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275435 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275444 4823 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275452 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275461 4823 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275470 4823 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275480 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275489 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275497 4823 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275507 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275516 4823 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275524 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.275533 4823 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.280917 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-ovn-kubernetes\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.281939 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38298bb6-9157-4c18-94a2-fedec7e95a4c-cni-binary-copy\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.281999 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6c32331c-bb6a-475a-847f-40b25351917e-host\") pod \"node-ca-7wn7x\" (UID: \"6c32331c-bb6a-475a-847f-40b25351917e\") " pod="openshift-image-registry/node-ca-7wn7x" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.282042 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-cnibin\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.282073 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-systemd-units\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.282102 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7322ac31-c6c4-43b9-8b3c-60ff84742545-rootfs\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.282131 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-var-lib-cni-bin\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.282540 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38298bb6-9157-4c18-94a2-fedec7e95a4c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.282579 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-ovn\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.282605 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-log-socket\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.282630 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-netd\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.282659 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-bin\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.282699 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-var-lib-kubelet\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.283087 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-slash\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.283129 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-etc-openvswitch\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.286429 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.286622 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-system-cni-dir\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.286661 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-etc-kubernetes\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.286712 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-os-release\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.286761 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-conf-dir\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.286796 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-node-log\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.286830 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7748c732-655d-42c5-9dee-f7c69315471b-hosts-file\") pod \"node-resolver-z4b65\" (UID: \"7748c732-655d-42c5-9dee-f7c69315471b\") " pod="openshift-dns/node-resolver-z4b65" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.286859 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-netns\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.286852 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-openvswitch\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.287237 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-socket-dir-parent\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.287378 4823 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.287454 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs podName:316f49ed-2e3c-42b7-a133-a36c3074810e nodeName:}" failed. No retries permitted until 2026-03-20 08:58:39.787428516 +0000 UTC m=+86.473324890 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs") pod "network-metrics-daemon-jzzk2" (UID: "316f49ed-2e3c-42b7-a133-a36c3074810e") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.287766 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-var-lib-openvswitch\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.288036 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-os-release\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.288062 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.288152 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-cnibin\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.288222 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6c32331c-bb6a-475a-847f-40b25351917e-serviceca\") pod \"node-ca-7wn7x\" (UID: \"6c32331c-bb6a-475a-847f-40b25351917e\") " pod="openshift-image-registry/node-ca-7wn7x" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.288377 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.288425 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-hostroot\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.288456 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38298bb6-9157-4c18-94a2-fedec7e95a4c-system-cni-dir\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.288683 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-config\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289122 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7322ac31-c6c4-43b9-8b3c-60ff84742545-mcd-auth-proxy-config\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289151 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-script-lib\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289203 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d33f9d63-34d7-4722-95d5-13bdcb32b99a-cni-binary-copy\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289225 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289271 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-run-multus-certs\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289307 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289307 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-systemd\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289366 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-run-k8s-cni-cncf-io\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289374 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-kubelet\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289406 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-run-netns\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289422 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-host-var-lib-cni-multus\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289451 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-cni-dir\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289694 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d33f9d63-34d7-4722-95d5-13bdcb32b99a-multus-daemon-config\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.289927 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.291320 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-env-overrides\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.300854 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.300885 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.301082 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.301152 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.301162 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.301179 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.301209 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.305164 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zf2d\" (UniqueName: \"kubernetes.io/projected/d33f9d63-34d7-4722-95d5-13bdcb32b99a-kube-api-access-2zf2d\") pod \"multus-9vsj7\" (UID: \"d33f9d63-34d7-4722-95d5-13bdcb32b99a\") " pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.306263 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovn-node-metrics-cert\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.308398 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7322ac31-c6c4-43b9-8b3c-60ff84742545-proxy-tls\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.311715 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgm2f\" (UniqueName: \"kubernetes.io/projected/7322ac31-c6c4-43b9-8b3c-60ff84742545-kube-api-access-cgm2f\") pod \"machine-config-daemon-snzb2\" (UID: \"7322ac31-c6c4-43b9-8b3c-60ff84742545\") " pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.312539 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgn6c\" (UniqueName: \"kubernetes.io/projected/6c32331c-bb6a-475a-847f-40b25351917e-kube-api-access-hgn6c\") pod \"node-ca-7wn7x\" (UID: \"6c32331c-bb6a-475a-847f-40b25351917e\") " pod="openshift-image-registry/node-ca-7wn7x" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.313008 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwkjs\" (UniqueName: \"kubernetes.io/projected/1bed1862-5bcd-46fe-bc84-2dbab416086b-kube-api-access-mwkjs\") pod \"ovnkube-node-2x54k\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.314890 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvgxr\" (UniqueName: \"kubernetes.io/projected/7748c732-655d-42c5-9dee-f7c69315471b-kube-api-access-tvgxr\") pod \"node-resolver-z4b65\" (UID: \"7748c732-655d-42c5-9dee-f7c69315471b\") " pod="openshift-dns/node-resolver-z4b65" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.314891 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dllqz\" (UniqueName: \"kubernetes.io/projected/74dc9b8c-5cfe-4336-863d-75c2a0ce028b-kube-api-access-dllqz\") pod \"ovnkube-control-plane-749d76644c-cx7cq\" (UID: \"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.315094 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4wm2\" (UniqueName: \"kubernetes.io/projected/316f49ed-2e3c-42b7-a133-a36c3074810e-kube-api-access-p4wm2\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.315133 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gfvd\" (UniqueName: \"kubernetes.io/projected/38298bb6-9157-4c18-94a2-fedec7e95a4c-kube-api-access-2gfvd\") pod \"multus-additional-cni-plugins-fb4mg\" (UID: \"38298bb6-9157-4c18-94a2-fedec7e95a4c\") " pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.323340 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.334832 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.378173 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.386470 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.393481 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.396392 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ -f "/env/_master" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: set -o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: source "/env/_master" Mar 20 08:58:39 crc kubenswrapper[4823]: set +o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 20 08:58:39 crc kubenswrapper[4823]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 20 08:58:39 crc kubenswrapper[4823]: ho_enable="--enable-hybrid-overlay" Mar 20 08:58:39 crc kubenswrapper[4823]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 20 08:58:39 crc kubenswrapper[4823]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 20 08:58:39 crc kubenswrapper[4823]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 20 08:58:39 crc kubenswrapper[4823]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 08:58:39 crc kubenswrapper[4823]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 20 08:58:39 crc kubenswrapper[4823]: --webhook-host=127.0.0.1 \ Mar 20 08:58:39 crc kubenswrapper[4823]: --webhook-port=9743 \ Mar 20 08:58:39 crc kubenswrapper[4823]: ${ho_enable} \ Mar 20 08:58:39 crc kubenswrapper[4823]: --enable-interconnect \ Mar 20 08:58:39 crc kubenswrapper[4823]: --disable-approver \ Mar 20 08:58:39 crc kubenswrapper[4823]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 20 08:58:39 crc kubenswrapper[4823]: --wait-for-kubernetes-api=200s \ Mar 20 08:58:39 crc kubenswrapper[4823]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 20 08:58:39 crc kubenswrapper[4823]: --loglevel="${LOGLEVEL}" Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.401584 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7wn7x" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.404952 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ -f "/env/_master" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: set -o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: source "/env/_master" Mar 20 08:58:39 crc kubenswrapper[4823]: set +o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 20 08:58:39 crc kubenswrapper[4823]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 08:58:39 crc kubenswrapper[4823]: --disable-webhook \ Mar 20 08:58:39 crc kubenswrapper[4823]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 20 08:58:39 crc kubenswrapper[4823]: --loglevel="${LOGLEVEL}" Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.406371 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.407076 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 20 08:58:39 crc kubenswrapper[4823]: set -o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: source /etc/kubernetes/apiserver-url.env Mar 20 08:58:39 crc kubenswrapper[4823]: else Mar 20 08:58:39 crc kubenswrapper[4823]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 20 08:58:39 crc kubenswrapper[4823]: exit 1 Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.408366 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.409774 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9vsj7" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.411665 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.411723 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.411733 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.411761 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.411775 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4823]: W0320 08:58:39.415552 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-1b677b027b8562009a30ff94c6dc740b267a0091c95646c0981bc06ed6cf0056 WatchSource:0}: Error finding container 1b677b027b8562009a30ff94c6dc740b267a0091c95646c0981bc06ed6cf0056: Status 404 returned error can't find the container with id 1b677b027b8562009a30ff94c6dc740b267a0091c95646c0981bc06ed6cf0056 Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.419208 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.421727 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.421791 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" Mar 20 08:58:39 crc kubenswrapper[4823]: W0320 08:58:39.425718 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c32331c_bb6a_475a_847f_40b25351917e.slice/crio-e2dc2d73361bb5a1b38fa2bd47aef79798ca5af3f929d939080e881786a2fa9d WatchSource:0}: Error finding container e2dc2d73361bb5a1b38fa2bd47aef79798ca5af3f929d939080e881786a2fa9d: Status 404 returned error can't find the container with id e2dc2d73361bb5a1b38fa2bd47aef79798ca5af3f929d939080e881786a2fa9d Mar 20 08:58:39 crc kubenswrapper[4823]: W0320 08:58:39.427650 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd33f9d63_34d7_4722_95d5_13bdcb32b99a.slice/crio-a9ffa00281a77a48f1d59c58bed4128b401eea726af571d52b992a49562a1b37 WatchSource:0}: Error finding container a9ffa00281a77a48f1d59c58bed4128b401eea726af571d52b992a49562a1b37: Status 404 returned error can't find the container with id a9ffa00281a77a48f1d59c58bed4128b401eea726af571d52b992a49562a1b37 Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.429268 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.432715 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Mar 20 08:58:39 crc kubenswrapper[4823]: while [ true ]; Mar 20 08:58:39 crc kubenswrapper[4823]: do Mar 20 08:58:39 crc kubenswrapper[4823]: for f in $(ls /tmp/serviceca); do Mar 20 08:58:39 crc kubenswrapper[4823]: echo $f Mar 20 08:58:39 crc kubenswrapper[4823]: ca_file_path="/tmp/serviceca/${f}" Mar 20 08:58:39 crc kubenswrapper[4823]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Mar 20 08:58:39 crc kubenswrapper[4823]: reg_dir_path="/etc/docker/certs.d/${f}" Mar 20 08:58:39 crc kubenswrapper[4823]: if [ -e "${reg_dir_path}" ]; then Mar 20 08:58:39 crc kubenswrapper[4823]: cp -u $ca_file_path $reg_dir_path/ca.crt Mar 20 08:58:39 crc kubenswrapper[4823]: else Mar 20 08:58:39 crc kubenswrapper[4823]: mkdir $reg_dir_path Mar 20 08:58:39 crc kubenswrapper[4823]: cp $ca_file_path $reg_dir_path/ca.crt Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: for d in $(ls /etc/docker/certs.d); do Mar 20 08:58:39 crc kubenswrapper[4823]: echo $d Mar 20 08:58:39 crc kubenswrapper[4823]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Mar 20 08:58:39 crc kubenswrapper[4823]: reg_conf_path="/tmp/serviceca/${dp}" Mar 20 08:58:39 crc kubenswrapper[4823]: if [ ! -e "${reg_conf_path}" ]; then Mar 20 08:58:39 crc kubenswrapper[4823]: rm -rf /etc/docker/certs.d/$d Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: sleep 60 & wait ${!} Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hgn6c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-7wn7x_openshift-image-registry(6c32331c-bb6a-475a-847f-40b25351917e): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.432836 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Mar 20 08:58:39 crc kubenswrapper[4823]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Mar 20 08:58:39 crc kubenswrapper[4823]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2zf2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-9vsj7_openshift-multus(d33f9d63-34d7-4722-95d5-13bdcb32b99a): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.433997 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-7wn7x" podUID="6c32331c-bb6a-475a-847f-40b25351917e" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.434019 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-9vsj7" podUID="d33f9d63-34d7-4722-95d5-13bdcb32b99a" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.439819 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.443474 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"696fc9c01008a9502ffaf586c66e7dd78d7e038eb72d9c7613d552c2b9884bff"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.445157 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"813a571f9239e1e53dd26d208c964959b3ef7d86bce6645059367f73dde17dab"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.448531 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9vsj7" event={"ID":"d33f9d63-34d7-4722-95d5-13bdcb32b99a","Type":"ContainerStarted","Data":"a9ffa00281a77a48f1d59c58bed4128b401eea726af571d52b992a49562a1b37"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.450024 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7wn7x" event={"ID":"6c32331c-bb6a-475a-847f-40b25351917e","Type":"ContainerStarted","Data":"e2dc2d73361bb5a1b38fa2bd47aef79798ca5af3f929d939080e881786a2fa9d"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.450827 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1b677b027b8562009a30ff94c6dc740b267a0091c95646c0981bc06ed6cf0056"} Mar 20 08:58:39 crc kubenswrapper[4823]: W0320 08:58:39.453811 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38298bb6_9157_4c18_94a2_fedec7e95a4c.slice/crio-518698a7f29c210e4cdaf16352fe285ddee441a6d8eb50a529a7c9f683fc8961 WatchSource:0}: Error finding container 518698a7f29c210e4cdaf16352fe285ddee441a6d8eb50a529a7c9f683fc8961: Status 404 returned error can't find the container with id 518698a7f29c210e4cdaf16352fe285ddee441a6d8eb50a529a7c9f683fc8961 Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.454273 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.456333 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Mar 20 08:58:39 crc kubenswrapper[4823]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Mar 20 08:58:39 crc kubenswrapper[4823]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2zf2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-9vsj7_openshift-multus(d33f9d63-34d7-4722-95d5-13bdcb32b99a): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.457194 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 20 08:58:39 crc kubenswrapper[4823]: set -o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: source /etc/kubernetes/apiserver-url.env Mar 20 08:58:39 crc kubenswrapper[4823]: else Mar 20 08:58:39 crc kubenswrapper[4823]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 20 08:58:39 crc kubenswrapper[4823]: exit 1 Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.457306 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Mar 20 08:58:39 crc kubenswrapper[4823]: while [ true ]; Mar 20 08:58:39 crc kubenswrapper[4823]: do Mar 20 08:58:39 crc kubenswrapper[4823]: for f in $(ls /tmp/serviceca); do Mar 20 08:58:39 crc kubenswrapper[4823]: echo $f Mar 20 08:58:39 crc kubenswrapper[4823]: ca_file_path="/tmp/serviceca/${f}" Mar 20 08:58:39 crc kubenswrapper[4823]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Mar 20 08:58:39 crc kubenswrapper[4823]: reg_dir_path="/etc/docker/certs.d/${f}" Mar 20 08:58:39 crc kubenswrapper[4823]: if [ -e "${reg_dir_path}" ]; then Mar 20 08:58:39 crc kubenswrapper[4823]: cp -u $ca_file_path $reg_dir_path/ca.crt Mar 20 08:58:39 crc kubenswrapper[4823]: else Mar 20 08:58:39 crc kubenswrapper[4823]: mkdir $reg_dir_path Mar 20 08:58:39 crc kubenswrapper[4823]: cp $ca_file_path $reg_dir_path/ca.crt Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: for d in $(ls /etc/docker/certs.d); do Mar 20 08:58:39 crc kubenswrapper[4823]: echo $d Mar 20 08:58:39 crc kubenswrapper[4823]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Mar 20 08:58:39 crc kubenswrapper[4823]: reg_conf_path="/tmp/serviceca/${dp}" Mar 20 08:58:39 crc kubenswrapper[4823]: if [ ! -e "${reg_conf_path}" ]; then Mar 20 08:58:39 crc kubenswrapper[4823]: rm -rf /etc/docker/certs.d/$d Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: sleep 60 & wait ${!} Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hgn6c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-7wn7x_openshift-image-registry(6c32331c-bb6a-475a-847f-40b25351917e): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.457388 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-9vsj7" podUID="d33f9d63-34d7-4722-95d5-13bdcb32b99a" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.457396 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2gfvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-fb4mg_openshift-multus(38298bb6-9157-4c18-94a2-fedec7e95a4c): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.457427 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: W0320 08:58:39.457793 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7322ac31_c6c4_43b9_8b3c_60ff84742545.slice/crio-305be5097b4cc63b319f63a75297654d65610a26da7dada9295360139c4ec09b WatchSource:0}: Error finding container 305be5097b4cc63b319f63a75297654d65610a26da7dada9295360139c4ec09b: Status 404 returned error can't find the container with id 305be5097b4cc63b319f63a75297654d65610a26da7dada9295360139c4ec09b Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.458311 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.458332 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ -f "/env/_master" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: set -o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: source "/env/_master" Mar 20 08:58:39 crc kubenswrapper[4823]: set +o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 20 08:58:39 crc kubenswrapper[4823]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 20 08:58:39 crc kubenswrapper[4823]: ho_enable="--enable-hybrid-overlay" Mar 20 08:58:39 crc kubenswrapper[4823]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 20 08:58:39 crc kubenswrapper[4823]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 20 08:58:39 crc kubenswrapper[4823]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 20 08:58:39 crc kubenswrapper[4823]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 08:58:39 crc kubenswrapper[4823]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 20 08:58:39 crc kubenswrapper[4823]: --webhook-host=127.0.0.1 \ Mar 20 08:58:39 crc kubenswrapper[4823]: --webhook-port=9743 \ Mar 20 08:58:39 crc kubenswrapper[4823]: ${ho_enable} \ Mar 20 08:58:39 crc kubenswrapper[4823]: --enable-interconnect \ Mar 20 08:58:39 crc kubenswrapper[4823]: --disable-approver \ Mar 20 08:58:39 crc kubenswrapper[4823]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 20 08:58:39 crc kubenswrapper[4823]: --wait-for-kubernetes-api=200s \ Mar 20 08:58:39 crc kubenswrapper[4823]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 20 08:58:39 crc kubenswrapper[4823]: --loglevel="${LOGLEVEL}" Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.458356 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-7wn7x" podUID="6c32331c-bb6a-475a-847f-40b25351917e" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.458474 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.458490 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" podUID="38298bb6-9157-4c18-94a2-fedec7e95a4c" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.458697 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.460530 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ -f "/env/_master" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: set -o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: source "/env/_master" Mar 20 08:58:39 crc kubenswrapper[4823]: set +o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 20 08:58:39 crc kubenswrapper[4823]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 08:58:39 crc kubenswrapper[4823]: --disable-webhook \ Mar 20 08:58:39 crc kubenswrapper[4823]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 20 08:58:39 crc kubenswrapper[4823]: --loglevel="${LOGLEVEL}" Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.461544 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cgm2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.461610 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.464067 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cgm2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: W0320 08:58:39.465323 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74dc9b8c_5cfe_4336_863d_75c2a0ce028b.slice/crio-0e3fc8e71d4b231d5b7913c3eee9f1f7c8d02b4a9a7ee081ab25b21f21638476 WatchSource:0}: Error finding container 0e3fc8e71d4b231d5b7913c3eee9f1f7c8d02b4a9a7ee081ab25b21f21638476: Status 404 returned error can't find the container with id 0e3fc8e71d4b231d5b7913c3eee9f1f7c8d02b4a9a7ee081ab25b21f21638476 Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.465395 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.467090 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.470842 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[/bin/bash -c #!/bin/bash Mar 20 08:58:39 crc kubenswrapper[4823]: set -euo pipefail Mar 20 08:58:39 crc kubenswrapper[4823]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Mar 20 08:58:39 crc kubenswrapper[4823]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Mar 20 08:58:39 crc kubenswrapper[4823]: # As the secret mount is optional we must wait for the files to be present. Mar 20 08:58:39 crc kubenswrapper[4823]: # The service is created in monitor.yaml and this is created in sdn.yaml. Mar 20 08:58:39 crc kubenswrapper[4823]: TS=$(date +%s) Mar 20 08:58:39 crc kubenswrapper[4823]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Mar 20 08:58:39 crc kubenswrapper[4823]: HAS_LOGGED_INFO=0 Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: log_missing_certs(){ Mar 20 08:58:39 crc kubenswrapper[4823]: CUR_TS=$(date +%s) Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Mar 20 08:58:39 crc kubenswrapper[4823]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Mar 20 08:58:39 crc kubenswrapper[4823]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Mar 20 08:58:39 crc kubenswrapper[4823]: HAS_LOGGED_INFO=1 Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: } Mar 20 08:58:39 crc kubenswrapper[4823]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Mar 20 08:58:39 crc kubenswrapper[4823]: log_missing_certs Mar 20 08:58:39 crc kubenswrapper[4823]: sleep 5 Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Mar 20 08:58:39 crc kubenswrapper[4823]: exec /usr/bin/kube-rbac-proxy \ Mar 20 08:58:39 crc kubenswrapper[4823]: --logtostderr \ Mar 20 08:58:39 crc kubenswrapper[4823]: --secure-listen-address=:9108 \ Mar 20 08:58:39 crc kubenswrapper[4823]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Mar 20 08:58:39 crc kubenswrapper[4823]: --upstream=http://127.0.0.1:29108/ \ Mar 20 08:58:39 crc kubenswrapper[4823]: --tls-private-key-file=${TLS_PK} \ Mar 20 08:58:39 crc kubenswrapper[4823]: --tls-cert-file=${TLS_CERT} Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dllqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-cx7cq_openshift-ovn-kubernetes(74dc9b8c-5cfe-4336-863d-75c2a0ce028b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.473615 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ -f "/env/_master" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: set -o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: source "/env/_master" Mar 20 08:58:39 crc kubenswrapper[4823]: set +o allexport Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: ovn_v4_join_subnet_opt= Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ "" != "" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: ovn_v6_join_subnet_opt= Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ "" != "" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: ovn_v4_transit_switch_subnet_opt= Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ "" != "" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: ovn_v6_transit_switch_subnet_opt= Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ "" != "" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: dns_name_resolver_enabled_flag= Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ "false" == "true" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: persistent_ips_enabled_flag= Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ "true" == "true" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: persistent_ips_enabled_flag="--enable-persistent-ips" Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: # This is needed so that converting clusters from GA to TP Mar 20 08:58:39 crc kubenswrapper[4823]: # will rollout control plane pods as well Mar 20 08:58:39 crc kubenswrapper[4823]: network_segmentation_enabled_flag= Mar 20 08:58:39 crc kubenswrapper[4823]: multi_network_enabled_flag= Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ "true" == "true" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: multi_network_enabled_flag="--enable-multi-network" Mar 20 08:58:39 crc kubenswrapper[4823]: network_segmentation_enabled_flag="--enable-network-segmentation" Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Mar 20 08:58:39 crc kubenswrapper[4823]: exec /usr/bin/ovnkube \ Mar 20 08:58:39 crc kubenswrapper[4823]: --enable-interconnect \ Mar 20 08:58:39 crc kubenswrapper[4823]: --init-cluster-manager "${K8S_NODE}" \ Mar 20 08:58:39 crc kubenswrapper[4823]: --config-file=/run/ovnkube-config/ovnkube.conf \ Mar 20 08:58:39 crc kubenswrapper[4823]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Mar 20 08:58:39 crc kubenswrapper[4823]: --metrics-bind-address "127.0.0.1:29108" \ Mar 20 08:58:39 crc kubenswrapper[4823]: --metrics-enable-pprof \ Mar 20 08:58:39 crc kubenswrapper[4823]: --metrics-enable-config-duration \ Mar 20 08:58:39 crc kubenswrapper[4823]: ${ovn_v4_join_subnet_opt} \ Mar 20 08:58:39 crc kubenswrapper[4823]: ${ovn_v6_join_subnet_opt} \ Mar 20 08:58:39 crc kubenswrapper[4823]: ${ovn_v4_transit_switch_subnet_opt} \ Mar 20 08:58:39 crc kubenswrapper[4823]: ${ovn_v6_transit_switch_subnet_opt} \ Mar 20 08:58:39 crc kubenswrapper[4823]: ${dns_name_resolver_enabled_flag} \ Mar 20 08:58:39 crc kubenswrapper[4823]: ${persistent_ips_enabled_flag} \ Mar 20 08:58:39 crc kubenswrapper[4823]: ${multi_network_enabled_flag} \ Mar 20 08:58:39 crc kubenswrapper[4823]: ${network_segmentation_enabled_flag} Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dllqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-cx7cq_openshift-ovn-kubernetes(74dc9b8c-5cfe-4336-863d-75c2a0ce028b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.475247 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" podUID="74dc9b8c-5cfe-4336-863d-75c2a0ce028b" Mar 20 08:58:39 crc kubenswrapper[4823]: W0320 08:58:39.481614 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bed1862_5bcd_46fe_bc84_2dbab416086b.slice/crio-e1e18bdec402b725c1c518a992d50ffec70ce8b27b41c8c41436593e42fbf51f WatchSource:0}: Error finding container e1e18bdec402b725c1c518a992d50ffec70ce8b27b41c8c41436593e42fbf51f: Status 404 returned error can't find the container with id e1e18bdec402b725c1c518a992d50ffec70ce8b27b41c8c41436593e42fbf51f Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.481691 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.485668 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Mar 20 08:58:39 crc kubenswrapper[4823]: apiVersion: v1 Mar 20 08:58:39 crc kubenswrapper[4823]: clusters: Mar 20 08:58:39 crc kubenswrapper[4823]: - cluster: Mar 20 08:58:39 crc kubenswrapper[4823]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Mar 20 08:58:39 crc kubenswrapper[4823]: server: https://api-int.crc.testing:6443 Mar 20 08:58:39 crc kubenswrapper[4823]: name: default-cluster Mar 20 08:58:39 crc kubenswrapper[4823]: contexts: Mar 20 08:58:39 crc kubenswrapper[4823]: - context: Mar 20 08:58:39 crc kubenswrapper[4823]: cluster: default-cluster Mar 20 08:58:39 crc kubenswrapper[4823]: namespace: default Mar 20 08:58:39 crc kubenswrapper[4823]: user: default-auth Mar 20 08:58:39 crc kubenswrapper[4823]: name: default-context Mar 20 08:58:39 crc kubenswrapper[4823]: current-context: default-context Mar 20 08:58:39 crc kubenswrapper[4823]: kind: Config Mar 20 08:58:39 crc kubenswrapper[4823]: preferences: {} Mar 20 08:58:39 crc kubenswrapper[4823]: users: Mar 20 08:58:39 crc kubenswrapper[4823]: - name: default-auth Mar 20 08:58:39 crc kubenswrapper[4823]: user: Mar 20 08:58:39 crc kubenswrapper[4823]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 20 08:58:39 crc kubenswrapper[4823]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 20 08:58:39 crc kubenswrapper[4823]: EOF Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mwkjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.486826 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.489113 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-z4b65" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.492550 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.500921 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: W0320 08:58:39.503933 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7748c732_655d_42c5_9dee_f7c69315471b.slice/crio-b6f85e278abed4c00790dfcb60ede9de1709b68cc99d5f15053562eeb9873fa5 WatchSource:0}: Error finding container b6f85e278abed4c00790dfcb60ede9de1709b68cc99d5f15053562eeb9873fa5: Status 404 returned error can't find the container with id b6f85e278abed4c00790dfcb60ede9de1709b68cc99d5f15053562eeb9873fa5 Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.506822 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:39 crc kubenswrapper[4823]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Mar 20 08:58:39 crc kubenswrapper[4823]: set -uo pipefail Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Mar 20 08:58:39 crc kubenswrapper[4823]: HOSTS_FILE="/etc/hosts" Mar 20 08:58:39 crc kubenswrapper[4823]: TEMP_FILE="/etc/hosts.tmp" Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: IFS=', ' read -r -a services <<< "${SERVICES}" Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: # Make a temporary file with the old hosts file's attributes. Mar 20 08:58:39 crc kubenswrapper[4823]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Mar 20 08:58:39 crc kubenswrapper[4823]: echo "Failed to preserve hosts file. Exiting." Mar 20 08:58:39 crc kubenswrapper[4823]: exit 1 Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: while true; do Mar 20 08:58:39 crc kubenswrapper[4823]: declare -A svc_ips Mar 20 08:58:39 crc kubenswrapper[4823]: for svc in "${services[@]}"; do Mar 20 08:58:39 crc kubenswrapper[4823]: # Fetch service IP from cluster dns if present. We make several tries Mar 20 08:58:39 crc kubenswrapper[4823]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Mar 20 08:58:39 crc kubenswrapper[4823]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Mar 20 08:58:39 crc kubenswrapper[4823]: # support UDP loadbalancers and require reaching DNS through TCP. Mar 20 08:58:39 crc kubenswrapper[4823]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 08:58:39 crc kubenswrapper[4823]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 08:58:39 crc kubenswrapper[4823]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 08:58:39 crc kubenswrapper[4823]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Mar 20 08:58:39 crc kubenswrapper[4823]: for i in ${!cmds[*]} Mar 20 08:58:39 crc kubenswrapper[4823]: do Mar 20 08:58:39 crc kubenswrapper[4823]: ips=($(eval "${cmds[i]}")) Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: svc_ips["${svc}"]="${ips[@]}" Mar 20 08:58:39 crc kubenswrapper[4823]: break Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: # Update /etc/hosts only if we get valid service IPs Mar 20 08:58:39 crc kubenswrapper[4823]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Mar 20 08:58:39 crc kubenswrapper[4823]: # Stale entries could exist in /etc/hosts if the service is deleted Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ -n "${svc_ips[*]-}" ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Mar 20 08:58:39 crc kubenswrapper[4823]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Mar 20 08:58:39 crc kubenswrapper[4823]: # Only continue rebuilding the hosts entries if its original content is preserved Mar 20 08:58:39 crc kubenswrapper[4823]: sleep 60 & wait Mar 20 08:58:39 crc kubenswrapper[4823]: continue Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: # Append resolver entries for services Mar 20 08:58:39 crc kubenswrapper[4823]: rc=0 Mar 20 08:58:39 crc kubenswrapper[4823]: for svc in "${!svc_ips[@]}"; do Mar 20 08:58:39 crc kubenswrapper[4823]: for ip in ${svc_ips[${svc}]}; do Mar 20 08:58:39 crc kubenswrapper[4823]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: if [[ $rc -ne 0 ]]; then Mar 20 08:58:39 crc kubenswrapper[4823]: sleep 60 & wait Mar 20 08:58:39 crc kubenswrapper[4823]: continue Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: Mar 20 08:58:39 crc kubenswrapper[4823]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Mar 20 08:58:39 crc kubenswrapper[4823]: # Replace /etc/hosts with our modified version if needed Mar 20 08:58:39 crc kubenswrapper[4823]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Mar 20 08:58:39 crc kubenswrapper[4823]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Mar 20 08:58:39 crc kubenswrapper[4823]: fi Mar 20 08:58:39 crc kubenswrapper[4823]: sleep 60 & wait Mar 20 08:58:39 crc kubenswrapper[4823]: unset svc_ips Mar 20 08:58:39 crc kubenswrapper[4823]: done Mar 20 08:58:39 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tvgxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-z4b65_openshift-dns(7748c732-655d-42c5-9dee-f7c69315471b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:39 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.508230 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-z4b65" podUID="7748c732-655d-42c5-9dee-f7c69315471b" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.511967 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.516554 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.516603 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.516617 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.516635 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.516647 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.521120 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.530883 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.540194 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.549064 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.563646 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.572925 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.584475 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.596371 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.611344 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.619221 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.619264 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.619275 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.619292 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.619844 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.621486 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.636047 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.645254 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.654855 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.670475 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.681011 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.681205 4823 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.681374 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:40.681335846 +0000 UTC m=+87.367232390 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.689694 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.700347 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.711764 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.722714 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.722785 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.722833 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.722852 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.722905 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.724463 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.735415 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.744293 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.754983 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.782445 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.782600 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.782653 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:58:40.782620129 +0000 UTC m=+87.468516503 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.782720 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.782825 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.782841 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.782896 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.782898 4823 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.782919 4823 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.782953 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.782976 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:40.78296478 +0000 UTC m=+87.468861154 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.782978 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.783000 4823 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.783017 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:40.782987501 +0000 UTC m=+87.468883905 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.783053 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:40.783035183 +0000 UTC m=+87.468931757 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.793633 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.825966 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.826011 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.826022 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.826037 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.826056 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.884120 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.884299 4823 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: E0320 08:58:39.884412 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs podName:316f49ed-2e3c-42b7-a133-a36c3074810e nodeName:}" failed. No retries permitted until 2026-03-20 08:58:40.884380087 +0000 UTC m=+87.570276501 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs") pod "network-metrics-daemon-jzzk2" (UID: "316f49ed-2e3c-42b7-a133-a36c3074810e") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.928972 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.929042 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.929059 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.929083 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4823]: I0320 08:58:39.929102 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.032066 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.032135 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.032153 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.032187 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.032216 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.103326 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.104105 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.105278 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.105966 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.106973 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.107491 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.108124 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.109087 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.109689 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.110631 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.111178 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.112262 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.112793 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.113289 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.114196 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.114723 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.115661 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.116089 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.116625 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.117645 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.118130 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.119132 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.119561 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.120578 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.120999 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.121573 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.122604 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.123081 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.124022 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.124473 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.125415 4823 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.125513 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.127218 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.128081 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.128478 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.130084 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.130722 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.131610 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.132259 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.134332 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.135645 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.135892 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.135960 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.135975 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.136011 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.136027 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.138133 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.139473 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.141522 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.142528 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.144492 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.145632 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.147795 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.148308 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.148809 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.149339 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.150191 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.151026 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.151582 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.239290 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.239373 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.239400 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.239429 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.239451 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.343654 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.343734 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.343788 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.343820 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.343841 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.447452 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.447498 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.447510 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.447527 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.447538 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.454897 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-z4b65" event={"ID":"7748c732-655d-42c5-9dee-f7c69315471b","Type":"ContainerStarted","Data":"b6f85e278abed4c00790dfcb60ede9de1709b68cc99d5f15053562eeb9873fa5"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.455935 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"e1e18bdec402b725c1c518a992d50ffec70ce8b27b41c8c41436593e42fbf51f"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.457252 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" event={"ID":"74dc9b8c-5cfe-4336-863d-75c2a0ce028b","Type":"ContainerStarted","Data":"0e3fc8e71d4b231d5b7913c3eee9f1f7c8d02b4a9a7ee081ab25b21f21638476"} Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.457360 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:40 crc kubenswrapper[4823]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Mar 20 08:58:40 crc kubenswrapper[4823]: apiVersion: v1 Mar 20 08:58:40 crc kubenswrapper[4823]: clusters: Mar 20 08:58:40 crc kubenswrapper[4823]: - cluster: Mar 20 08:58:40 crc kubenswrapper[4823]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Mar 20 08:58:40 crc kubenswrapper[4823]: server: https://api-int.crc.testing:6443 Mar 20 08:58:40 crc kubenswrapper[4823]: name: default-cluster Mar 20 08:58:40 crc kubenswrapper[4823]: contexts: Mar 20 08:58:40 crc kubenswrapper[4823]: - context: Mar 20 08:58:40 crc kubenswrapper[4823]: cluster: default-cluster Mar 20 08:58:40 crc kubenswrapper[4823]: namespace: default Mar 20 08:58:40 crc kubenswrapper[4823]: user: default-auth Mar 20 08:58:40 crc kubenswrapper[4823]: name: default-context Mar 20 08:58:40 crc kubenswrapper[4823]: current-context: default-context Mar 20 08:58:40 crc kubenswrapper[4823]: kind: Config Mar 20 08:58:40 crc kubenswrapper[4823]: preferences: {} Mar 20 08:58:40 crc kubenswrapper[4823]: users: Mar 20 08:58:40 crc kubenswrapper[4823]: - name: default-auth Mar 20 08:58:40 crc kubenswrapper[4823]: user: Mar 20 08:58:40 crc kubenswrapper[4823]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 20 08:58:40 crc kubenswrapper[4823]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 20 08:58:40 crc kubenswrapper[4823]: EOF Mar 20 08:58:40 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mwkjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:40 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.457494 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:40 crc kubenswrapper[4823]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Mar 20 08:58:40 crc kubenswrapper[4823]: set -uo pipefail Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Mar 20 08:58:40 crc kubenswrapper[4823]: HOSTS_FILE="/etc/hosts" Mar 20 08:58:40 crc kubenswrapper[4823]: TEMP_FILE="/etc/hosts.tmp" Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: IFS=', ' read -r -a services <<< "${SERVICES}" Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: # Make a temporary file with the old hosts file's attributes. Mar 20 08:58:40 crc kubenswrapper[4823]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Mar 20 08:58:40 crc kubenswrapper[4823]: echo "Failed to preserve hosts file. Exiting." Mar 20 08:58:40 crc kubenswrapper[4823]: exit 1 Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: while true; do Mar 20 08:58:40 crc kubenswrapper[4823]: declare -A svc_ips Mar 20 08:58:40 crc kubenswrapper[4823]: for svc in "${services[@]}"; do Mar 20 08:58:40 crc kubenswrapper[4823]: # Fetch service IP from cluster dns if present. We make several tries Mar 20 08:58:40 crc kubenswrapper[4823]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Mar 20 08:58:40 crc kubenswrapper[4823]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Mar 20 08:58:40 crc kubenswrapper[4823]: # support UDP loadbalancers and require reaching DNS through TCP. Mar 20 08:58:40 crc kubenswrapper[4823]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 08:58:40 crc kubenswrapper[4823]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 08:58:40 crc kubenswrapper[4823]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 08:58:40 crc kubenswrapper[4823]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Mar 20 08:58:40 crc kubenswrapper[4823]: for i in ${!cmds[*]} Mar 20 08:58:40 crc kubenswrapper[4823]: do Mar 20 08:58:40 crc kubenswrapper[4823]: ips=($(eval "${cmds[i]}")) Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: svc_ips["${svc}"]="${ips[@]}" Mar 20 08:58:40 crc kubenswrapper[4823]: break Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: done Mar 20 08:58:40 crc kubenswrapper[4823]: done Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: # Update /etc/hosts only if we get valid service IPs Mar 20 08:58:40 crc kubenswrapper[4823]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Mar 20 08:58:40 crc kubenswrapper[4823]: # Stale entries could exist in /etc/hosts if the service is deleted Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ -n "${svc_ips[*]-}" ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Mar 20 08:58:40 crc kubenswrapper[4823]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Mar 20 08:58:40 crc kubenswrapper[4823]: # Only continue rebuilding the hosts entries if its original content is preserved Mar 20 08:58:40 crc kubenswrapper[4823]: sleep 60 & wait Mar 20 08:58:40 crc kubenswrapper[4823]: continue Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: # Append resolver entries for services Mar 20 08:58:40 crc kubenswrapper[4823]: rc=0 Mar 20 08:58:40 crc kubenswrapper[4823]: for svc in "${!svc_ips[@]}"; do Mar 20 08:58:40 crc kubenswrapper[4823]: for ip in ${svc_ips[${svc}]}; do Mar 20 08:58:40 crc kubenswrapper[4823]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Mar 20 08:58:40 crc kubenswrapper[4823]: done Mar 20 08:58:40 crc kubenswrapper[4823]: done Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ $rc -ne 0 ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: sleep 60 & wait Mar 20 08:58:40 crc kubenswrapper[4823]: continue Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Mar 20 08:58:40 crc kubenswrapper[4823]: # Replace /etc/hosts with our modified version if needed Mar 20 08:58:40 crc kubenswrapper[4823]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Mar 20 08:58:40 crc kubenswrapper[4823]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: sleep 60 & wait Mar 20 08:58:40 crc kubenswrapper[4823]: unset svc_ips Mar 20 08:58:40 crc kubenswrapper[4823]: done Mar 20 08:58:40 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tvgxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-z4b65_openshift-dns(7748c732-655d-42c5-9dee-f7c69315471b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:40 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.458510 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.459267 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-z4b65" podUID="7748c732-655d-42c5-9dee-f7c69315471b" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.459544 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"305be5097b4cc63b319f63a75297654d65610a26da7dada9295360139c4ec09b"} Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.459949 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:40 crc kubenswrapper[4823]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[/bin/bash -c #!/bin/bash Mar 20 08:58:40 crc kubenswrapper[4823]: set -euo pipefail Mar 20 08:58:40 crc kubenswrapper[4823]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Mar 20 08:58:40 crc kubenswrapper[4823]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Mar 20 08:58:40 crc kubenswrapper[4823]: # As the secret mount is optional we must wait for the files to be present. Mar 20 08:58:40 crc kubenswrapper[4823]: # The service is created in monitor.yaml and this is created in sdn.yaml. Mar 20 08:58:40 crc kubenswrapper[4823]: TS=$(date +%s) Mar 20 08:58:40 crc kubenswrapper[4823]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Mar 20 08:58:40 crc kubenswrapper[4823]: HAS_LOGGED_INFO=0 Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: log_missing_certs(){ Mar 20 08:58:40 crc kubenswrapper[4823]: CUR_TS=$(date +%s) Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Mar 20 08:58:40 crc kubenswrapper[4823]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Mar 20 08:58:40 crc kubenswrapper[4823]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Mar 20 08:58:40 crc kubenswrapper[4823]: HAS_LOGGED_INFO=1 Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: } Mar 20 08:58:40 crc kubenswrapper[4823]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Mar 20 08:58:40 crc kubenswrapper[4823]: log_missing_certs Mar 20 08:58:40 crc kubenswrapper[4823]: sleep 5 Mar 20 08:58:40 crc kubenswrapper[4823]: done Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Mar 20 08:58:40 crc kubenswrapper[4823]: exec /usr/bin/kube-rbac-proxy \ Mar 20 08:58:40 crc kubenswrapper[4823]: --logtostderr \ Mar 20 08:58:40 crc kubenswrapper[4823]: --secure-listen-address=:9108 \ Mar 20 08:58:40 crc kubenswrapper[4823]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Mar 20 08:58:40 crc kubenswrapper[4823]: --upstream=http://127.0.0.1:29108/ \ Mar 20 08:58:40 crc kubenswrapper[4823]: --tls-private-key-file=${TLS_PK} \ Mar 20 08:58:40 crc kubenswrapper[4823]: --tls-cert-file=${TLS_CERT} Mar 20 08:58:40 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dllqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-cx7cq_openshift-ovn-kubernetes(74dc9b8c-5cfe-4336-863d-75c2a0ce028b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:40 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.460795 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cgm2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.460940 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" event={"ID":"38298bb6-9157-4c18-94a2-fedec7e95a4c","Type":"ContainerStarted","Data":"518698a7f29c210e4cdaf16352fe285ddee441a6d8eb50a529a7c9f683fc8961"} Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.462012 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2gfvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-fb4mg_openshift-multus(38298bb6-9157-4c18-94a2-fedec7e95a4c): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.462047 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:40 crc kubenswrapper[4823]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ -f "/env/_master" ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: set -o allexport Mar 20 08:58:40 crc kubenswrapper[4823]: source "/env/_master" Mar 20 08:58:40 crc kubenswrapper[4823]: set +o allexport Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: ovn_v4_join_subnet_opt= Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ "" != "" ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: ovn_v6_join_subnet_opt= Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ "" != "" ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: ovn_v4_transit_switch_subnet_opt= Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ "" != "" ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: ovn_v6_transit_switch_subnet_opt= Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ "" != "" ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: dns_name_resolver_enabled_flag= Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ "false" == "true" ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: persistent_ips_enabled_flag= Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ "true" == "true" ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: persistent_ips_enabled_flag="--enable-persistent-ips" Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: # This is needed so that converting clusters from GA to TP Mar 20 08:58:40 crc kubenswrapper[4823]: # will rollout control plane pods as well Mar 20 08:58:40 crc kubenswrapper[4823]: network_segmentation_enabled_flag= Mar 20 08:58:40 crc kubenswrapper[4823]: multi_network_enabled_flag= Mar 20 08:58:40 crc kubenswrapper[4823]: if [[ "true" == "true" ]]; then Mar 20 08:58:40 crc kubenswrapper[4823]: multi_network_enabled_flag="--enable-multi-network" Mar 20 08:58:40 crc kubenswrapper[4823]: network_segmentation_enabled_flag="--enable-network-segmentation" Mar 20 08:58:40 crc kubenswrapper[4823]: fi Mar 20 08:58:40 crc kubenswrapper[4823]: Mar 20 08:58:40 crc kubenswrapper[4823]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Mar 20 08:58:40 crc kubenswrapper[4823]: exec /usr/bin/ovnkube \ Mar 20 08:58:40 crc kubenswrapper[4823]: --enable-interconnect \ Mar 20 08:58:40 crc kubenswrapper[4823]: --init-cluster-manager "${K8S_NODE}" \ Mar 20 08:58:40 crc kubenswrapper[4823]: --config-file=/run/ovnkube-config/ovnkube.conf \ Mar 20 08:58:40 crc kubenswrapper[4823]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Mar 20 08:58:40 crc kubenswrapper[4823]: --metrics-bind-address "127.0.0.1:29108" \ Mar 20 08:58:40 crc kubenswrapper[4823]: --metrics-enable-pprof \ Mar 20 08:58:40 crc kubenswrapper[4823]: --metrics-enable-config-duration \ Mar 20 08:58:40 crc kubenswrapper[4823]: ${ovn_v4_join_subnet_opt} \ Mar 20 08:58:40 crc kubenswrapper[4823]: ${ovn_v6_join_subnet_opt} \ Mar 20 08:58:40 crc kubenswrapper[4823]: ${ovn_v4_transit_switch_subnet_opt} \ Mar 20 08:58:40 crc kubenswrapper[4823]: ${ovn_v6_transit_switch_subnet_opt} \ Mar 20 08:58:40 crc kubenswrapper[4823]: ${dns_name_resolver_enabled_flag} \ Mar 20 08:58:40 crc kubenswrapper[4823]: ${persistent_ips_enabled_flag} \ Mar 20 08:58:40 crc kubenswrapper[4823]: ${multi_network_enabled_flag} \ Mar 20 08:58:40 crc kubenswrapper[4823]: ${network_segmentation_enabled_flag} Mar 20 08:58:40 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dllqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-cx7cq_openshift-ovn-kubernetes(74dc9b8c-5cfe-4336-863d-75c2a0ce028b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:40 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.462766 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cgm2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.463830 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" podUID="38298bb6-9157-4c18-94a2-fedec7e95a4c" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.463886 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.464332 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" podUID="74dc9b8c-5cfe-4336-863d-75c2a0ce028b" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.471602 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.484717 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.503631 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.514588 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.525348 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.541011 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.550044 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.550097 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.550110 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.550134 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.550148 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.551166 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.562706 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.572623 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.581708 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.594305 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.612288 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.633887 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.644363 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.653225 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.653274 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.653284 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.653311 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.653322 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.654733 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.666673 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.676641 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.691150 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.695885 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.696041 4823 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.696127 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:42.696106945 +0000 UTC m=+89.382003469 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.701813 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.708919 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.720050 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.734336 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.748262 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.756133 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.756288 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.756369 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.756450 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.756523 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.769188 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.796585 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.797065 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.797288 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.797374 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.797407 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.797423 4823 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.797612 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.797703 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.797792 4823 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.797382 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:58:42.797341715 +0000 UTC m=+89.483238119 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.797923 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:42.797903743 +0000 UTC m=+89.483800117 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.797943 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.797989 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:42.797981966 +0000 UTC m=+89.483878340 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.798062 4823 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.798090 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:42.798082829 +0000 UTC m=+89.483979213 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.799794 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.841347 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.858319 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.858358 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.858369 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.858386 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.858399 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.874219 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.898696 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.898955 4823 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:40 crc kubenswrapper[4823]: E0320 08:58:40.900226 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs podName:316f49ed-2e3c-42b7-a133-a36c3074810e nodeName:}" failed. No retries permitted until 2026-03-20 08:58:42.900197899 +0000 UTC m=+89.586094303 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs") pod "network-metrics-daemon-jzzk2" (UID: "316f49ed-2e3c-42b7-a133-a36c3074810e") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.914214 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.961102 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.961179 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.961190 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.961207 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4823]: I0320 08:58:40.961219 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.064583 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.065072 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.065197 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.065350 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.065467 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.096461 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.096551 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:41 crc kubenswrapper[4823]: E0320 08:58:41.096660 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.097017 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:41 crc kubenswrapper[4823]: E0320 08:58:41.097162 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.097215 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:41 crc kubenswrapper[4823]: E0320 08:58:41.097331 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:41 crc kubenswrapper[4823]: E0320 08:58:41.097419 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.168313 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.168387 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.168409 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.168439 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.168461 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.270854 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.270893 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.270906 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.270924 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.270939 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.373428 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.373527 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.373558 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.373587 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.373606 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.476845 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.476927 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.476952 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.476983 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.477006 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.579647 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.579693 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.579707 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.579726 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.579780 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.683013 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.683077 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.683095 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.683122 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.683139 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.786345 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.786404 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.786423 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.786447 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.786463 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.889702 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.889766 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.889779 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.889797 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.889810 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.993620 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.993666 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.993677 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.993693 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4823]: I0320 08:58:41.993706 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.096543 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.096593 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.096601 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.096612 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.096622 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.204271 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.204345 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.204369 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.204398 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.204416 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.307435 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.307498 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.307517 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.307547 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.307574 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.410997 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.411051 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.411062 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.411104 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.411120 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.513318 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.513372 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.513383 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.513400 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.513410 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.617504 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.617571 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.617584 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.617604 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.617618 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.721439 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.721484 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.721493 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.721507 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.721519 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.721851 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.722124 4823 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.722308 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:46.722265732 +0000 UTC m=+93.408162276 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.823216 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.823599 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:58:46.823556495 +0000 UTC m=+93.509452899 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.823808 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.823895 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.823969 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.824066 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.824093 4823 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.824109 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.824148 4823 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.824185 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:46.824160743 +0000 UTC m=+93.510057147 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.824259 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:46.824203235 +0000 UTC m=+93.510099759 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.824368 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.824415 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.824441 4823 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.824543 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.824590 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.824609 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.824631 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.824647 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.824562 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:46.824538945 +0000 UTC m=+93.510435479 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.924702 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.924920 4823 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:42 crc kubenswrapper[4823]: E0320 08:58:42.925023 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs podName:316f49ed-2e3c-42b7-a133-a36c3074810e nodeName:}" failed. No retries permitted until 2026-03-20 08:58:46.924992131 +0000 UTC m=+93.610888655 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs") pod "network-metrics-daemon-jzzk2" (UID: "316f49ed-2e3c-42b7-a133-a36c3074810e") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.927388 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.927444 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.927458 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.927478 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4823]: I0320 08:58:42.927490 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.031164 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.031237 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.031261 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.031291 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.031316 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.063676 4823 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.097219 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.097253 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.097265 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.097224 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:43 crc kubenswrapper[4823]: E0320 08:58:43.097374 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:43 crc kubenswrapper[4823]: E0320 08:58:43.097521 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:43 crc kubenswrapper[4823]: E0320 08:58:43.097614 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:58:43 crc kubenswrapper[4823]: E0320 08:58:43.097823 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.134018 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.134080 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.134096 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.134121 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.134137 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.237015 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.237082 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.237102 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.237120 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.237132 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.339922 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.339964 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.339978 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.339994 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.340005 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.443110 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.443177 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.443195 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.443220 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.443240 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.545558 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.545586 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.545595 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.545607 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.545615 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.648370 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.648406 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.648415 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.648429 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.648439 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.751628 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.751669 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.751683 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.751699 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.751711 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.854530 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.854567 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.854575 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.854590 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.854599 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.957833 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.957903 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.957922 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.957950 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4823]: I0320 08:58:43.957969 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.060401 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.060475 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.060492 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.060517 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.060535 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.118275 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.131377 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.135542 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.135606 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.135623 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.135645 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.135659 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.141660 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: E0320 08:58:44.148903 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.152782 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.152822 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.152845 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.152884 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.152922 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.155215 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: E0320 08:58:44.166476 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.167998 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.170721 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.170795 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.170808 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.170835 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.170850 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.178433 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: E0320 08:58:44.180909 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.184611 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.184643 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.184656 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.184678 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.184691 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.186489 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: E0320 08:58:44.194655 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.199143 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.199181 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.199191 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.199208 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.199222 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.206985 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: E0320 08:58:44.214352 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: E0320 08:58:44.214521 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.216609 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.216636 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.216674 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.216693 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.216709 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.217430 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.238033 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.248949 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.257843 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.267094 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.282662 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.319841 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.319920 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.319932 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.319954 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.319965 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.422497 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.422527 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.422535 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.422548 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.422559 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.525440 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.525510 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.525529 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.525558 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.525578 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.627928 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.627975 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.627988 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.628005 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.628017 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.730791 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.730849 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.730872 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.730898 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.730920 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.833448 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.833877 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.834149 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.834341 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.834546 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.937471 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.937546 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.937558 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.937573 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4823]: I0320 08:58:44.937584 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.039975 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.040057 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.040081 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.040113 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.040135 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.097274 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.097400 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:45 crc kubenswrapper[4823]: E0320 08:58:45.097522 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.097539 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.097599 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:45 crc kubenswrapper[4823]: E0320 08:58:45.098041 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:58:45 crc kubenswrapper[4823]: E0320 08:58:45.098230 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:45 crc kubenswrapper[4823]: E0320 08:58:45.098434 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.143173 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.143221 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.143233 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.143252 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.143273 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.245247 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.245307 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.245327 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.245353 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.245370 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.349130 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.349186 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.349205 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.349229 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.349246 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.452238 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.452285 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.452299 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.452317 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.452332 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.554907 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.554980 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.555007 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.555036 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.555058 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.657487 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.657555 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.657580 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.657608 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.657634 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.760765 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.760807 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.760819 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.760833 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.760846 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.863984 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.864032 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.864045 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.864060 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.864072 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.967427 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.967477 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.967500 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.967523 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4823]: I0320 08:58:45.967540 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.070370 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.070481 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.070502 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.070525 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.070545 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.110388 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.173919 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.173979 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.173997 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.174024 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.174044 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.276855 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.278184 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.278236 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.278269 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.278303 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.382650 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.382718 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.382773 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.382808 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.382829 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.485534 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.485575 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.485586 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.485607 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.485618 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.588553 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.588889 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.589008 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.589129 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.589226 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.692154 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.692226 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.692240 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.692257 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.692324 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.768147 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.768271 4823 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.768332 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:54.768312677 +0000 UTC m=+101.454209051 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.795567 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.795618 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.795635 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.795659 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.795676 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.869225 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.869525 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:58:54.869484265 +0000 UTC m=+101.555380659 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.869653 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.869836 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.869902 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.869992 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.870048 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.870066 4823 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.870108 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.870154 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:54.870129876 +0000 UTC m=+101.556026260 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.870161 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.870190 4823 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.870234 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:54.870218948 +0000 UTC m=+101.556115522 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.870111 4823 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.870281 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:54.87026884 +0000 UTC m=+101.556165224 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.897591 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.897656 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.897675 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.897697 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.897713 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4823]: I0320 08:58:46.971296 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.971546 4823 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:46 crc kubenswrapper[4823]: E0320 08:58:46.971830 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs podName:316f49ed-2e3c-42b7-a133-a36c3074810e nodeName:}" failed. No retries permitted until 2026-03-20 08:58:54.971802831 +0000 UTC m=+101.657699215 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs") pod "network-metrics-daemon-jzzk2" (UID: "316f49ed-2e3c-42b7-a133-a36c3074810e") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.001009 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.001076 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.001095 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.001120 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.001137 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.096723 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.096784 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.096849 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.096908 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:47 crc kubenswrapper[4823]: E0320 08:58:47.096906 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:47 crc kubenswrapper[4823]: E0320 08:58:47.097034 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:47 crc kubenswrapper[4823]: E0320 08:58:47.097182 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:47 crc kubenswrapper[4823]: E0320 08:58:47.097442 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.104579 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.104657 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.104673 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.104707 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.104727 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.207790 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.207865 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.207879 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.207897 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.207910 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.310586 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.310637 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.310651 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.310668 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.310681 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.413352 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.413729 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.413858 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.413955 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.414054 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.517481 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.517884 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.517978 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.518050 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.518123 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.621104 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.621514 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.621649 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.621906 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.622056 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.724922 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.724988 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.725012 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.725047 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.725071 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.829236 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.829318 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.829342 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.829372 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.829392 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.932388 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.932462 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.932480 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.932510 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4823]: I0320 08:58:47.932528 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.035527 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.035580 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.035598 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.035615 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.035626 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.138708 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.138838 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.139155 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.139435 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.139493 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.242330 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.242389 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.242400 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.242414 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.242424 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.345276 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.345356 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.345380 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.345413 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.345437 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.448359 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.448439 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.448460 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.448484 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.448502 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.552195 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.552594 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.552699 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.552861 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.552936 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.656203 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.656270 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.656321 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.656345 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.656362 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.766785 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.766852 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.766870 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.766895 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.766913 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.869623 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.869697 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.869712 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.869735 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.869775 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.973437 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.973514 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.973543 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.973579 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4823]: I0320 08:58:48.973605 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.076857 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.076915 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.076930 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.076952 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.076967 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.096525 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.096631 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.096724 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.096555 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:49 crc kubenswrapper[4823]: E0320 08:58:49.096810 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:49 crc kubenswrapper[4823]: E0320 08:58:49.096959 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:58:49 crc kubenswrapper[4823]: E0320 08:58:49.097464 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:49 crc kubenswrapper[4823]: E0320 08:58:49.097583 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.112924 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.114026 4823 scope.go:117] "RemoveContainer" containerID="6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099" Mar 20 08:58:49 crc kubenswrapper[4823]: E0320 08:58:49.114300 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.179732 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.179829 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.179842 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.179863 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.179879 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.282509 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.282563 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.282575 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.282594 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.282608 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.386151 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.386212 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.386234 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.386261 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.386280 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.489647 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.489825 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.489856 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.489956 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.490034 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.492008 4823 scope.go:117] "RemoveContainer" containerID="6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099" Mar 20 08:58:49 crc kubenswrapper[4823]: E0320 08:58:49.492264 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.593273 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.593332 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.593348 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.593370 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.593389 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.695914 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.695972 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.695984 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.696004 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.696018 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.799607 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.799674 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.799692 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.799723 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.799779 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.902702 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.902817 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.902834 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.902852 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4823]: I0320 08:58:49.902863 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.006235 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.006292 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.006310 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.006332 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.006351 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.108835 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.108864 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.108873 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.108885 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.108894 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.111078 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.211610 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.211649 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.211661 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.211676 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.211687 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.314072 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.314117 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.314129 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.314147 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.314160 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.417634 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.417696 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.417721 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.417764 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.417791 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.520273 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.520324 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.520337 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.520354 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.520366 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.623805 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.623853 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.623864 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.623882 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.623895 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.726520 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.726582 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.726595 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.726615 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.726626 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.829140 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.829205 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.829225 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.829255 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.829274 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.932123 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.932188 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.932205 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.932228 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4823]: I0320 08:58:50.932246 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.035158 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.035213 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.035224 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.035245 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.035258 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.097004 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.097099 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.097241 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.097685 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:51 crc kubenswrapper[4823]: E0320 08:58:51.097928 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:51 crc kubenswrapper[4823]: E0320 08:58:51.098084 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:51 crc kubenswrapper[4823]: E0320 08:58:51.098640 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:51 crc kubenswrapper[4823]: E0320 08:58:51.098767 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.138053 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.138099 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.138112 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.138132 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.138144 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.240473 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.240517 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.240528 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.240546 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.240558 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.343644 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.343695 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.343712 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.343761 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.343782 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.447196 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.447252 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.447263 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.447284 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.447295 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.501032 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" event={"ID":"38298bb6-9157-4c18-94a2-fedec7e95a4c","Type":"ContainerStarted","Data":"c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.503730 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48" exitCode=0 Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.503814 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.517563 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.530453 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.550247 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.550590 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.550691 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.550805 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.550891 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.554296 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.567944 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.588849 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.600815 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.611949 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.628985 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.642339 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.661421 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.661525 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.661749 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.661779 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.661994 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.662022 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.675552 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.685734 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.700569 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.715885 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.735343 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.754932 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.764453 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.767094 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.767145 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.767162 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.767187 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.767204 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.778416 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.788944 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.805959 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.829101 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.842987 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.853624 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.866305 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.870162 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.870212 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.870228 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.870248 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.870266 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.876473 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.890559 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.899385 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.910663 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.926055 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.943173 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.971132 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.972928 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.972960 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.972972 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.972989 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.973002 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.980928 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:51 crc kubenswrapper[4823]: I0320 08:58:51.992535 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.007400 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.076493 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.076525 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.076535 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.076550 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.076561 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.179373 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.179405 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.179415 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.179430 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.179441 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.282490 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.282549 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.282561 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.282578 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.282592 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.385433 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.385474 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.385485 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.385500 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.385510 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.488891 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.488957 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.488972 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.488999 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.489015 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.508442 4823 generic.go:334] "Generic (PLEG): container finished" podID="38298bb6-9157-4c18-94a2-fedec7e95a4c" containerID="c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808" exitCode=0 Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.508538 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" event={"ID":"38298bb6-9157-4c18-94a2-fedec7e95a4c","Type":"ContainerDied","Data":"c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.515124 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.515586 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.515669 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.515734 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.515817 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.515878 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.525463 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.537843 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.553527 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.565107 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.575130 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.586359 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.591631 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.591668 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.591679 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.591693 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.591703 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.598875 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.611942 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.623668 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.651631 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.661663 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.670877 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.687843 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.694639 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.694693 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.694707 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.694757 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.694772 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.697230 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.714076 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.728974 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.751405 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.798555 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.798605 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.798619 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.798639 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.798652 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.901972 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.902461 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.902674 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.903050 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4823]: I0320 08:58:52.903249 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.006064 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.006363 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.006501 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.006615 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.006707 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.097203 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.097275 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:53 crc kubenswrapper[4823]: E0320 08:58:53.097383 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.097439 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.097623 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:53 crc kubenswrapper[4823]: E0320 08:58:53.097813 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:53 crc kubenswrapper[4823]: E0320 08:58:53.098088 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:53 crc kubenswrapper[4823]: E0320 08:58:53.098245 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.114440 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.114843 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.115008 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.115118 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.115240 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.217310 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.217427 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.217446 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.217472 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.217489 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.319552 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.319680 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.319769 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.319874 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.320008 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.456202 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.456236 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.456246 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.456265 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.456275 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.521087 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.523349 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9vsj7" event={"ID":"d33f9d63-34d7-4722-95d5-13bdcb32b99a","Type":"ContainerStarted","Data":"05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.525127 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.525189 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.533636 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.547115 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.558040 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.558095 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.558114 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.558138 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.558156 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.562868 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.580509 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.591091 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.606205 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.619176 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.637799 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.650611 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.660789 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.660840 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.660852 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.660870 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.660883 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.666792 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.678654 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.693272 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.705904 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.731321 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.749883 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.763512 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.763556 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.763567 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.763582 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.763594 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.766761 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.777612 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.788518 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.797573 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.815412 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.824572 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.831457 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.843784 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.853695 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.862103 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.865849 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.865896 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.865911 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.865927 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.865936 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.875944 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.896213 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.906641 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.916257 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.928111 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.939834 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.953974 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.968603 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.968656 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.968667 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.968682 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.968693 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.969586 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:53 crc kubenswrapper[4823]: I0320 08:58:53.983893 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.071948 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.072026 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.072045 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.072075 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.072094 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.119729 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.140066 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.156101 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.175024 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.175065 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.175077 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.175095 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.175108 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.177776 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.194466 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.210984 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.227308 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.240768 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.255863 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.270019 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.278158 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.278239 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.278256 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.278298 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.278311 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.289829 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.304429 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.316943 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.331784 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.353037 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.368601 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.380781 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.380825 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.380839 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.380863 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.380876 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.385535 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.484167 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.484227 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.484243 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.484267 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.484285 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.513196 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.513246 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.513265 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.513292 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.513308 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.531827 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.534755 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.536329 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.536371 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.536384 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.536400 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.536413 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.538138 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.540811 4823 generic.go:334] "Generic (PLEG): container finished" podID="38298bb6-9157-4c18-94a2-fedec7e95a4c" containerID="dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e" exitCode=0 Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.541424 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" event={"ID":"38298bb6-9157-4c18-94a2-fedec7e95a4c","Type":"ContainerDied","Data":"dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e"} Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.549959 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.554653 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.555367 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.555407 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.555420 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.555439 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.555452 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.571663 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.573698 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.581029 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.581061 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.581075 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.581092 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.581104 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.585592 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.592851 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.596968 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.597007 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.597023 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.597042 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.597055 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.600556 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.609358 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.609802 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.612055 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.612100 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.612112 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.612132 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.612145 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.617229 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.630394 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.642612 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.656855 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.667585 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.677915 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.686864 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.698935 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.710788 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.715433 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.715468 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.715478 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.715494 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.715503 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.728729 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.740890 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.754965 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.772712 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.787170 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.798412 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.818217 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.818887 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.818961 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.818987 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.819017 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.819040 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.836280 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.854612 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.869056 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.869206 4823 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.869295 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:10.869272498 +0000 UTC m=+117.555168882 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.870186 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.883269 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.895670 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.907625 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.921285 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.922219 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.922252 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.922262 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.922282 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.922295 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.934168 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.944235 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.962552 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.969813 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.970016 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:59:10.969982802 +0000 UTC m=+117.655879336 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.970061 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.970089 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.970120 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.970274 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.970299 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.970310 4823 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.970313 4823 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.970320 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.970364 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.970371 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:10.970354934 +0000 UTC m=+117.656251308 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.970383 4823 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.970391 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:10.970383275 +0000 UTC m=+117.656279649 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:54 crc kubenswrapper[4823]: E0320 08:58:54.970715 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:10.970706275 +0000 UTC m=+117.656602649 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.978392 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:54 crc kubenswrapper[4823]: I0320 08:58:54.991099 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.012193 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.024884 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.025242 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.025274 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.025283 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.025300 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.025310 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.071414 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:55 crc kubenswrapper[4823]: E0320 08:58:55.071587 4823 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:55 crc kubenswrapper[4823]: E0320 08:58:55.071653 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs podName:316f49ed-2e3c-42b7-a133-a36c3074810e nodeName:}" failed. No retries permitted until 2026-03-20 08:59:11.071636267 +0000 UTC m=+117.757532641 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs") pod "network-metrics-daemon-jzzk2" (UID: "316f49ed-2e3c-42b7-a133-a36c3074810e") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.097166 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.097218 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.097181 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.097331 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:55 crc kubenswrapper[4823]: E0320 08:58:55.097718 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:55 crc kubenswrapper[4823]: E0320 08:58:55.098015 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:55 crc kubenswrapper[4823]: E0320 08:58:55.098076 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:55 crc kubenswrapper[4823]: E0320 08:58:55.098161 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.135285 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.135321 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.135460 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.135480 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.135495 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.238673 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.238733 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.238769 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.238795 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.238809 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.345931 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.346347 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.346363 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.346386 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.346408 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.454313 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.454368 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.454387 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.454410 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.454427 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.547726 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7wn7x" event={"ID":"6c32331c-bb6a-475a-847f-40b25351917e","Type":"ContainerStarted","Data":"00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.550405 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-z4b65" event={"ID":"7748c732-655d-42c5-9dee-f7c69315471b","Type":"ContainerStarted","Data":"2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.556618 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.558065 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.558121 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.558143 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.558167 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.558186 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.560311 4823 generic.go:334] "Generic (PLEG): container finished" podID="38298bb6-9157-4c18-94a2-fedec7e95a4c" containerID="27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766" exitCode=0 Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.560350 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" event={"ID":"38298bb6-9157-4c18-94a2-fedec7e95a4c","Type":"ContainerDied","Data":"27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.568443 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.597115 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.609948 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.619895 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.639929 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.661456 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.674063 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.687691 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.693320 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.693378 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.693393 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.693413 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.693428 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.711917 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.730987 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.742807 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.758256 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.773652 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.785492 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.795628 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.795666 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.795678 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.795694 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.795703 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.797963 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.810814 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.821632 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.832335 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.863255 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.883643 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.898807 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.898841 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.898851 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.898866 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.898877 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.921642 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.933326 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.950262 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.963301 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4823]: I0320 08:58:55.988625 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.001624 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.001704 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.001724 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.001796 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.001813 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.003558 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.036149 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.069243 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.104639 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.104718 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.104765 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.104787 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.104798 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.121230 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.153052 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.194417 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.208056 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.208110 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.208125 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.208143 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.208156 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.233088 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.272735 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.310942 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.311025 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.311039 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.311080 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.311092 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.313516 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.413816 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.413905 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.413916 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.413931 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.413942 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.518517 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.518660 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.518676 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.518697 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.518708 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.568668 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" event={"ID":"74dc9b8c-5cfe-4336-863d-75c2a0ce028b","Type":"ContainerStarted","Data":"7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.568727 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" event={"ID":"74dc9b8c-5cfe-4336-863d-75c2a0ce028b","Type":"ContainerStarted","Data":"8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.572628 4823 generic.go:334] "Generic (PLEG): container finished" podID="38298bb6-9157-4c18-94a2-fedec7e95a4c" containerID="948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c" exitCode=0 Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.572685 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" event={"ID":"38298bb6-9157-4c18-94a2-fedec7e95a4c","Type":"ContainerDied","Data":"948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.600143 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.618376 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.623580 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.623765 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.623776 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.623977 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.623992 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.638904 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.654817 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.667272 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.679548 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.695908 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.714587 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.728046 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.731640 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.731730 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.731772 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.731794 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.731812 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.743454 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.755676 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.791535 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.835949 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.836014 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.836030 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.836054 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.836067 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.837761 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.875869 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.924864 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.938763 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.938806 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.938816 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.938832 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.938844 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.959559 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:56 crc kubenswrapper[4823]: I0320 08:58:56.991720 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.031624 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.041118 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.041147 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.041158 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.041177 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.041190 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.076683 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.097064 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.097097 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.097212 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:57 crc kubenswrapper[4823]: E0320 08:58:57.097315 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.097328 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:57 crc kubenswrapper[4823]: E0320 08:58:57.097475 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:57 crc kubenswrapper[4823]: E0320 08:58:57.097622 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:57 crc kubenswrapper[4823]: E0320 08:58:57.097783 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.114874 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.143850 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.143880 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.143890 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.143904 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.143915 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.153668 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.194499 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.232839 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.246717 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.246773 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.246784 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.246800 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.246812 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.273566 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.314701 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.351849 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.351910 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.351930 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.351957 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.351976 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.354659 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.393819 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.437868 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.454525 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.454594 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.454609 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.454633 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.454649 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.470794 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.516867 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.553162 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.557596 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.557638 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.557656 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.557680 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.557699 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.583688 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.589362 4823 generic.go:334] "Generic (PLEG): container finished" podID="38298bb6-9157-4c18-94a2-fedec7e95a4c" containerID="056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659" exitCode=0 Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.589413 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" event={"ID":"38298bb6-9157-4c18-94a2-fedec7e95a4c","Type":"ContainerDied","Data":"056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.608529 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.639996 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.661784 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.661822 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.661831 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.661846 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.661857 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.674308 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.715122 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.753853 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.764724 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.764789 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.764799 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.764814 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.764853 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.802810 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.835415 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.867559 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.867599 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.867609 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.867626 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.867639 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.881212 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.914545 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.953865 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.969567 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.969601 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.969610 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.969624 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.969635 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4823]: I0320 08:58:57.995559 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.034203 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.073343 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.073423 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.073442 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.073472 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.073487 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.078952 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.117543 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.155260 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.176487 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.176561 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.176582 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.176609 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.176628 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.196598 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.242653 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.279482 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.279571 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.279591 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.279621 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.279650 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.281617 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.332271 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.361706 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.383818 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.383872 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.383886 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.383904 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.383916 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.486991 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.487052 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.487067 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.487087 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.487102 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.590864 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.590908 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.590922 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.590941 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.590954 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.598385 4823 generic.go:334] "Generic (PLEG): container finished" podID="38298bb6-9157-4c18-94a2-fedec7e95a4c" containerID="569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e" exitCode=0 Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.598519 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" event={"ID":"38298bb6-9157-4c18-94a2-fedec7e95a4c","Type":"ContainerDied","Data":"569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e"} Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.605382 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"a8e29a53c4cf9306c7b4cff09c90001a4915e57bf7ca80d318943e5e0591a42c"} Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.606416 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.606476 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.606501 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.621973 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.637336 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.637866 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.643106 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.657250 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.674417 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.689017 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.696237 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.696310 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.696334 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.696366 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.696390 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.702232 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.716095 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.727690 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.738823 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.753615 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.792992 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.803119 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.803157 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.803169 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.803185 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.803197 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.830809 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.870616 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.905763 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.905793 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.905801 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.905815 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.905824 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.914999 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.956798 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:58 crc kubenswrapper[4823]: I0320 08:58:58.999525 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.008785 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.008821 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.008834 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.008850 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.008871 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.035059 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.072454 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.096266 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.096329 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.096266 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:59 crc kubenswrapper[4823]: E0320 08:58:59.096391 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:59 crc kubenswrapper[4823]: E0320 08:58:59.096482 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.096496 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:58:59 crc kubenswrapper[4823]: E0320 08:58:59.096555 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:59 crc kubenswrapper[4823]: E0320 08:58:59.096648 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.112288 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.112436 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.112464 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.112473 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.112489 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.112499 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.161879 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e29a53c4cf9306c7b4cff09c90001a4915e57bf7ca80d318943e5e0591a42c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.190634 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.215138 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.215167 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.215175 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.215188 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.215198 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.230525 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.274165 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.316032 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.318330 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.318367 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.318384 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.318405 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.318420 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.353535 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.397914 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.420889 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.421248 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.421318 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.421398 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.421512 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.454125 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.485141 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.514874 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.524021 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.524062 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.524077 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.524094 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.524107 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.556208 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.594228 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.614345 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" event={"ID":"38298bb6-9157-4c18-94a2-fedec7e95a4c","Type":"ContainerStarted","Data":"f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174"} Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.626138 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.626186 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.626195 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.626213 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.626231 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.633759 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.672322 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.714664 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.729024 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.729074 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.729086 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.729103 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.729118 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.758356 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.804385 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.832097 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.832157 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.832170 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.832188 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.832522 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.835222 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.879074 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.922139 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.935881 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.935926 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.935936 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.935952 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.935963 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.952904 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:59 crc kubenswrapper[4823]: I0320 08:58:59.995486 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.033545 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.038796 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.038842 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.038857 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.038875 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.038889 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:00Z","lastTransitionTime":"2026-03-20T08:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.073379 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.115394 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.142019 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.142065 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.142077 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.142097 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.142110 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:00Z","lastTransitionTime":"2026-03-20T08:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.154704 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.203587 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e29a53c4cf9306c7b4cff09c90001a4915e57bf7ca80d318943e5e0591a42c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.232083 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.245075 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.245392 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.245506 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.245607 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.245727 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:00Z","lastTransitionTime":"2026-03-20T08:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.272277 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.316726 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.348905 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.349268 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.349470 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.349667 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.349879 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:00Z","lastTransitionTime":"2026-03-20T08:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.365127 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.398608 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.453321 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.453702 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.453882 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.454039 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.454204 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:00Z","lastTransitionTime":"2026-03-20T08:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.557772 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.557812 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.557821 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.557837 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.557848 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:00Z","lastTransitionTime":"2026-03-20T08:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.618892 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/0.log" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.622530 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="a8e29a53c4cf9306c7b4cff09c90001a4915e57bf7ca80d318943e5e0591a42c" exitCode=1 Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.622647 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"a8e29a53c4cf9306c7b4cff09c90001a4915e57bf7ca80d318943e5e0591a42c"} Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.623519 4823 scope.go:117] "RemoveContainer" containerID="a8e29a53c4cf9306c7b4cff09c90001a4915e57bf7ca80d318943e5e0591a42c" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.647806 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.662187 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.662563 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.662590 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.662613 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.662630 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:00Z","lastTransitionTime":"2026-03-20T08:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.674555 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.708618 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.722379 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.741486 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.758128 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.765962 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.766015 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.766027 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.766048 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.766062 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:00Z","lastTransitionTime":"2026-03-20T08:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.773451 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.783441 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.800765 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.814147 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.830724 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.869585 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.869630 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.869645 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.869666 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.869680 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:00Z","lastTransitionTime":"2026-03-20T08:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.872169 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.913045 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.954227 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.971907 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.971972 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.971987 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.972011 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.972026 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:00Z","lastTransitionTime":"2026-03-20T08:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:00 crc kubenswrapper[4823]: I0320 08:59:00.993422 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.043441 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e29a53c4cf9306c7b4cff09c90001a4915e57bf7ca80d318943e5e0591a42c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e29a53c4cf9306c7b4cff09c90001a4915e57bf7ca80d318943e5e0591a42c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:00Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:00.162710 6661 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:59:00.162763 6661 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:59:00.162800 6661 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:59:00.162817 6661 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:59:00.162813 6661 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:59:00.162845 6661 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:59:00.162847 6661 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:59:00.162861 6661 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:59:00.162876 6661 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0320 08:59:00.162878 6661 handler.go:208] Removed *v1.Node event handler 7\\\\nI0320 08:59:00.162895 6661 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:59:00.162950 6661 factory.go:656] Stopping watch factory\\\\nI0320 08:59:00.162969 6661 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:59:00.162979 6661 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:00.162951 6661 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0320 08:59:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.071695 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.075121 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.075152 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.075162 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.075177 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.075186 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:01Z","lastTransitionTime":"2026-03-20T08:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.096900 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.096939 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.096901 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:01 crc kubenswrapper[4823]: E0320 08:59:01.097027 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.097167 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:01 crc kubenswrapper[4823]: E0320 08:59:01.097300 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:01 crc kubenswrapper[4823]: E0320 08:59:01.097362 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:01 crc kubenswrapper[4823]: E0320 08:59:01.097454 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.177977 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.178023 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.178036 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.178057 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.178070 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:01Z","lastTransitionTime":"2026-03-20T08:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.281211 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.281259 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.281269 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.281286 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.281305 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:01Z","lastTransitionTime":"2026-03-20T08:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.383724 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.383784 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.383792 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.383810 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.383820 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:01Z","lastTransitionTime":"2026-03-20T08:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.486852 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.486907 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.486921 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.486939 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.486949 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:01Z","lastTransitionTime":"2026-03-20T08:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.589990 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.590055 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.590070 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.590096 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.590112 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:01Z","lastTransitionTime":"2026-03-20T08:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.630915 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/0.log" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.633788 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf"} Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.634396 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.647452 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.662257 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.674959 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.687137 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.693013 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.693046 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.693058 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.693072 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.693081 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:01Z","lastTransitionTime":"2026-03-20T08:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.701192 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.715646 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.728251 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.742664 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.766126 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e29a53c4cf9306c7b4cff09c90001a4915e57bf7ca80d318943e5e0591a42c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:00Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:00.162710 6661 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:59:00.162763 6661 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:59:00.162800 6661 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:59:00.162817 6661 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:59:00.162813 6661 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:59:00.162845 6661 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:59:00.162847 6661 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:59:00.162861 6661 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:59:00.162876 6661 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0320 08:59:00.162878 6661 handler.go:208] Removed *v1.Node event handler 7\\\\nI0320 08:59:00.162895 6661 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:59:00.162950 6661 factory.go:656] Stopping watch factory\\\\nI0320 08:59:00.162969 6661 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:59:00.162979 6661 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:00.162951 6661 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0320 08:59:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.778787 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.793616 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.795704 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.795747 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.795758 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.795773 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.795784 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:01Z","lastTransitionTime":"2026-03-20T08:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.812636 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.836009 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.850221 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.873508 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.899718 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.899778 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.899790 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.899806 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.899819 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:01Z","lastTransitionTime":"2026-03-20T08:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.907053 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4823]: I0320 08:59:01.923909 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:01Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.003434 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.003487 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.003499 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.003519 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.003531 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:02Z","lastTransitionTime":"2026-03-20T08:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.106851 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.106911 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.106922 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.106942 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.106955 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:02Z","lastTransitionTime":"2026-03-20T08:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.210926 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.211000 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.211016 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.211041 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.211081 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:02Z","lastTransitionTime":"2026-03-20T08:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.315376 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.315436 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.315455 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.315481 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.315498 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:02Z","lastTransitionTime":"2026-03-20T08:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.418894 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.419028 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.419054 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.419088 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.419110 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:02Z","lastTransitionTime":"2026-03-20T08:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.522507 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.522952 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.523150 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.523353 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.523553 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:02Z","lastTransitionTime":"2026-03-20T08:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.627174 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.627256 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.627285 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.627313 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.627334 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:02Z","lastTransitionTime":"2026-03-20T08:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.653335 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/1.log" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.654317 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/0.log" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.658761 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf" exitCode=1 Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.658807 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf"} Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.658858 4823 scope.go:117] "RemoveContainer" containerID="a8e29a53c4cf9306c7b4cff09c90001a4915e57bf7ca80d318943e5e0591a42c" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.659782 4823 scope.go:117] "RemoveContainer" containerID="916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf" Mar 20 08:59:02 crc kubenswrapper[4823]: E0320 08:59:02.659998 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.729657 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.729698 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.729708 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.729724 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:02 crc kubenswrapper[4823]: I0320 08:59:02.729752 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:02Z","lastTransitionTime":"2026-03-20T08:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.082644 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.082709 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.082727 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.082774 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.082788 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:03Z","lastTransitionTime":"2026-03-20T08:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.096524 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.096566 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.096641 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.096707 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:03 crc kubenswrapper[4823]: E0320 08:59:03.096865 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.097112 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: E0320 08:59:03.097359 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:03 crc kubenswrapper[4823]: E0320 08:59:03.097429 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:03 crc kubenswrapper[4823]: E0320 08:59:03.097483 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.114391 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.132001 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.145103 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.172411 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e29a53c4cf9306c7b4cff09c90001a4915e57bf7ca80d318943e5e0591a42c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:00Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:00.162710 6661 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:59:00.162763 6661 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:59:00.162800 6661 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:59:00.162817 6661 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:59:00.162813 6661 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:59:00.162845 6661 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:59:00.162847 6661 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:59:00.162861 6661 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:59:00.162876 6661 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0320 08:59:00.162878 6661 handler.go:208] Removed *v1.Node event handler 7\\\\nI0320 08:59:00.162895 6661 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:59:00.162950 6661 factory.go:656] Stopping watch factory\\\\nI0320 08:59:00.162969 6661 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:59:00.162979 6661 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:00.162951 6661 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0320 08:59:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:01Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.680919 6842 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:59:01.680927 6842 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:59:01.680932 6842 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:59:01.680939 6842 handler.go:208] Removed *v1.Node event handler 7\\\\nI0320 08:59:01.681072 6842 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0320 08:59:01.681284 6842 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681710 6842 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681769 6842 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.681789 6842 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.185870 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.185932 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.185942 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.185962 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.185973 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:03Z","lastTransitionTime":"2026-03-20T08:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.189135 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.201391 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.221960 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.239240 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.263051 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.284962 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.289488 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.289570 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.289586 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.289638 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.289656 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:03Z","lastTransitionTime":"2026-03-20T08:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.301619 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.321221 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.338865 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.358620 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.370661 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.385627 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.392862 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.392910 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.392926 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.392948 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.392963 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:03Z","lastTransitionTime":"2026-03-20T08:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.495488 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.495529 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.495540 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.495558 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.495568 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:03Z","lastTransitionTime":"2026-03-20T08:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.598367 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.598698 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.598785 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.598858 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.598927 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:03Z","lastTransitionTime":"2026-03-20T08:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.662952 4823 scope.go:117] "RemoveContainer" containerID="916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf" Mar 20 08:59:03 crc kubenswrapper[4823]: E0320 08:59:03.663151 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.702813 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.702861 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.702872 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.702891 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.702906 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:03Z","lastTransitionTime":"2026-03-20T08:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.805434 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.805489 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.805504 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.805526 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.805539 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:03Z","lastTransitionTime":"2026-03-20T08:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.908679 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.908726 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.908754 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.908777 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:03 crc kubenswrapper[4823]: I0320 08:59:03.908793 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:03Z","lastTransitionTime":"2026-03-20T08:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.011715 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.011781 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.011793 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.011809 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.011820 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.057449 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.072921 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.087682 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.097512 4823 scope.go:117] "RemoveContainer" containerID="6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.099675 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.117588 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.117724 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.117772 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.117781 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.117793 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.117804 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.132127 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.143771 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.159352 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.171344 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.184636 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.203458 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:01Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.680919 6842 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:59:01.680927 6842 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:59:01.680932 6842 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:59:01.680939 6842 handler.go:208] Removed *v1.Node event handler 7\\\\nI0320 08:59:01.681072 6842 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0320 08:59:01.681284 6842 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681710 6842 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681769 6842 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.681789 6842 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.215322 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.224095 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.224131 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.224142 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.224155 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.224166 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.234928 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.247894 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.268425 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.281590 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.296094 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.316166 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.328799 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.328843 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.328854 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.328878 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.328891 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.329497 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.349366 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.360194 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.373758 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.388884 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.400646 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.409783 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.423209 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.431838 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.431888 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.431900 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.431919 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.431937 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.435820 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.447679 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.457338 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.474492 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:01Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.680919 6842 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:59:01.680927 6842 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:59:01.680932 6842 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:59:01.680939 6842 handler.go:208] Removed *v1.Node event handler 7\\\\nI0320 08:59:01.681072 6842 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0320 08:59:01.681284 6842 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681710 6842 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681769 6842 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.681789 6842 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.488862 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.502464 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.518554 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.535786 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.535834 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.535845 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.535866 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.535878 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.536676 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.639035 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.639115 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.639137 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.639167 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.639188 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.668214 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/1.log" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.675257 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.677676 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.678496 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.699548 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.714155 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.731053 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.741671 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.741707 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.741722 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.741765 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.741781 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.748596 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.748646 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.748659 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.748678 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.748699 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.755399 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: E0320 08:59:04.769252 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.773207 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.773251 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.773266 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.773282 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.773296 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.782547 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: E0320 08:59:04.791294 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.795153 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.795186 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.795198 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.795217 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.795231 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.802220 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: E0320 08:59:04.813060 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.818111 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.818170 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.818183 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.818202 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.818217 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.820696 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: E0320 08:59:04.832222 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.836186 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.836239 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.836256 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.836284 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.836300 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.837226 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.849665 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: E0320 08:59:04.852035 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: E0320 08:59:04.852186 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.853902 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.853951 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.853964 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.853986 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.854000 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.859076 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.872542 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.887483 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.909616 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.932316 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.955561 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:01Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.680919 6842 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:59:01.680927 6842 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:59:01.680932 6842 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:59:01.680939 6842 handler.go:208] Removed *v1.Node event handler 7\\\\nI0320 08:59:01.681072 6842 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0320 08:59:01.681284 6842 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681710 6842 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681769 6842 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.681789 6842 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.956607 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.956663 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.956676 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.956701 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.956719 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:04Z","lastTransitionTime":"2026-03-20T08:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.971649 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4823]: I0320 08:59:04.987726 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.059377 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.059429 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.059441 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.059465 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.059478 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.096489 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.096548 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.096573 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:05 crc kubenswrapper[4823]: E0320 08:59:05.096622 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.096667 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:05 crc kubenswrapper[4823]: E0320 08:59:05.096824 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:05 crc kubenswrapper[4823]: E0320 08:59:05.097013 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:05 crc kubenswrapper[4823]: E0320 08:59:05.097103 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.162360 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.162414 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.162424 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.162445 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.162457 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.265554 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.265623 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.265636 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.265665 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.265682 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.369054 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.369111 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.369123 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.369147 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.369164 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.472253 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.472320 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.472333 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.472353 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.472592 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.575454 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.575553 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.575573 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.575599 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.575618 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.678213 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.678261 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.678271 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.678288 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.678300 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.782266 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.782357 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.782379 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.782408 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.782427 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.885722 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.885853 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.885879 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.885918 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.885946 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.988597 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.988684 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.988707 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.988734 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4823]: I0320 08:59:05.988781 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.096032 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.096108 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.096128 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.096154 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.096171 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:06Z","lastTransitionTime":"2026-03-20T08:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.199493 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.199566 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.199584 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.199610 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.199629 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:06Z","lastTransitionTime":"2026-03-20T08:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.302795 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.302864 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.302884 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.302931 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.302953 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:06Z","lastTransitionTime":"2026-03-20T08:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.406159 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.406222 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.406240 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.406262 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.406279 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:06Z","lastTransitionTime":"2026-03-20T08:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.509640 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.509718 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.509780 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.509814 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.509835 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:06Z","lastTransitionTime":"2026-03-20T08:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.612731 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.612842 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.612860 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.612883 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.612900 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:06Z","lastTransitionTime":"2026-03-20T08:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.715366 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.715432 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.715472 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.715500 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.715519 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:06Z","lastTransitionTime":"2026-03-20T08:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.818791 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.818858 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.818878 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.818904 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.818922 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:06Z","lastTransitionTime":"2026-03-20T08:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.922275 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.922331 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.922352 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.922376 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:06 crc kubenswrapper[4823]: I0320 08:59:06.922393 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:06Z","lastTransitionTime":"2026-03-20T08:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.025432 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.025504 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.025522 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.025546 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.025562 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:07Z","lastTransitionTime":"2026-03-20T08:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.097038 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.097089 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.097134 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.097097 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:07 crc kubenswrapper[4823]: E0320 08:59:07.097197 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:07 crc kubenswrapper[4823]: E0320 08:59:07.097298 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:07 crc kubenswrapper[4823]: E0320 08:59:07.097411 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:07 crc kubenswrapper[4823]: E0320 08:59:07.097506 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.130426 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.130462 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.130473 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.130489 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.130500 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:07Z","lastTransitionTime":"2026-03-20T08:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.234508 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.234580 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.234604 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.234634 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.234658 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:07Z","lastTransitionTime":"2026-03-20T08:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.338384 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.338428 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.338439 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.338453 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.338463 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:07Z","lastTransitionTime":"2026-03-20T08:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.441793 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.441843 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.441855 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.441875 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.441890 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:07Z","lastTransitionTime":"2026-03-20T08:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.544669 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.544716 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.544736 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.544798 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.544816 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:07Z","lastTransitionTime":"2026-03-20T08:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.649711 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.649813 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.649834 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.649862 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.649880 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:07Z","lastTransitionTime":"2026-03-20T08:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.753620 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.753672 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.753690 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.753713 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.753730 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:07Z","lastTransitionTime":"2026-03-20T08:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.857618 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.857715 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.857772 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.857811 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.857835 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:07Z","lastTransitionTime":"2026-03-20T08:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.962212 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.962650 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.962686 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.962722 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:07 crc kubenswrapper[4823]: I0320 08:59:07.962779 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:07Z","lastTransitionTime":"2026-03-20T08:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.065784 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.065854 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.065872 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.065901 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.065920 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:08Z","lastTransitionTime":"2026-03-20T08:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.168860 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.168926 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.168944 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.168970 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.168988 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:08Z","lastTransitionTime":"2026-03-20T08:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.272536 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.272603 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.272620 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.272653 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.272672 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:08Z","lastTransitionTime":"2026-03-20T08:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.376431 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.376521 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.376539 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.376565 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.376584 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:08Z","lastTransitionTime":"2026-03-20T08:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.478631 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.478669 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.478677 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.478692 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.478701 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:08Z","lastTransitionTime":"2026-03-20T08:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.581661 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.581706 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.581716 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.581730 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.581758 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:08Z","lastTransitionTime":"2026-03-20T08:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.683985 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.684054 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.684079 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.684104 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.684119 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:08Z","lastTransitionTime":"2026-03-20T08:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.787084 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.787131 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.787144 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.787160 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.787173 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:08Z","lastTransitionTime":"2026-03-20T08:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.891537 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.891586 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.891603 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.891622 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.891637 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:08Z","lastTransitionTime":"2026-03-20T08:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.995298 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.995363 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.995380 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.995398 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:08 crc kubenswrapper[4823]: I0320 08:59:08.995410 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:08Z","lastTransitionTime":"2026-03-20T08:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.096608 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.096635 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.096666 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:09 crc kubenswrapper[4823]: E0320 08:59:09.096789 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:09 crc kubenswrapper[4823]: E0320 08:59:09.096912 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.096941 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:09 crc kubenswrapper[4823]: E0320 08:59:09.097081 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:09 crc kubenswrapper[4823]: E0320 08:59:09.097143 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.101066 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.101176 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.101194 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.101216 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.101230 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:09Z","lastTransitionTime":"2026-03-20T08:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.203896 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.204020 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.204043 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.204068 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.204084 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:09Z","lastTransitionTime":"2026-03-20T08:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.306581 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.306639 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.306655 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.306680 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.306696 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:09Z","lastTransitionTime":"2026-03-20T08:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.410621 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.410680 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.410697 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.410719 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.410738 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:09Z","lastTransitionTime":"2026-03-20T08:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.513512 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.513576 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.513593 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.513616 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.513634 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:09Z","lastTransitionTime":"2026-03-20T08:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.617507 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.617572 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.617588 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.617613 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.617630 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:09Z","lastTransitionTime":"2026-03-20T08:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.721391 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.721469 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.721495 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.721519 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.721536 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:09Z","lastTransitionTime":"2026-03-20T08:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.825327 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.825400 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.825423 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.825451 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.825473 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:09Z","lastTransitionTime":"2026-03-20T08:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.929209 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.929302 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.929324 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.929353 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:09 crc kubenswrapper[4823]: I0320 08:59:09.929375 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:09Z","lastTransitionTime":"2026-03-20T08:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.032801 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.032865 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.032887 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.032912 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.032929 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:10Z","lastTransitionTime":"2026-03-20T08:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.112015 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.135365 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.135405 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.135421 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.135439 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.135454 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:10Z","lastTransitionTime":"2026-03-20T08:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.238722 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.238803 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.238821 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.238846 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.238863 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:10Z","lastTransitionTime":"2026-03-20T08:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.341995 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.342051 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.342069 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.342105 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.342122 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:10Z","lastTransitionTime":"2026-03-20T08:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.445956 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.446047 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.446077 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.446112 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.446138 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:10Z","lastTransitionTime":"2026-03-20T08:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.549498 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.549550 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.549561 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.549583 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.549595 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:10Z","lastTransitionTime":"2026-03-20T08:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.653478 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.653542 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.653560 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.653586 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.653604 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:10Z","lastTransitionTime":"2026-03-20T08:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.756332 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.756400 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.756413 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.756430 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.756442 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:10Z","lastTransitionTime":"2026-03-20T08:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.859808 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.859878 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.859905 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.859940 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.859964 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:10Z","lastTransitionTime":"2026-03-20T08:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.956642 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:10 crc kubenswrapper[4823]: E0320 08:59:10.956788 4823 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:59:10 crc kubenswrapper[4823]: E0320 08:59:10.956883 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:42.956860562 +0000 UTC m=+149.642756946 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.962347 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.962775 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.962933 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.963105 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:10 crc kubenswrapper[4823]: I0320 08:59:10.963322 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:10Z","lastTransitionTime":"2026-03-20T08:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.057954 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.058089 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.058122 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.058147 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.058274 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.058289 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.058298 4823 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.058340 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:43.058326919 +0000 UTC m=+149.744223293 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.058661 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:59:43.05865251 +0000 UTC m=+149.744548884 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.058705 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.058714 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.058721 4823 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.058758 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:43.058734473 +0000 UTC m=+149.744630847 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.058805 4823 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.058826 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:43.058820615 +0000 UTC m=+149.744716979 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.065677 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.065775 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.065798 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.065822 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.065838 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:11Z","lastTransitionTime":"2026-03-20T08:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.096992 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.097115 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.097273 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.097315 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.097540 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.097590 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.097623 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.097658 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.159843 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.160112 4823 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:59:11 crc kubenswrapper[4823]: E0320 08:59:11.160598 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs podName:316f49ed-2e3c-42b7-a133-a36c3074810e nodeName:}" failed. No retries permitted until 2026-03-20 08:59:43.160564483 +0000 UTC m=+149.846460887 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs") pod "network-metrics-daemon-jzzk2" (UID: "316f49ed-2e3c-42b7-a133-a36c3074810e") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.169002 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.169085 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.169113 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.169141 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.169160 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:11Z","lastTransitionTime":"2026-03-20T08:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.272134 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.272186 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.272202 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.272222 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.272238 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:11Z","lastTransitionTime":"2026-03-20T08:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.375202 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.375265 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.375283 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.375310 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.375331 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:11Z","lastTransitionTime":"2026-03-20T08:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.478249 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.478295 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.478306 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.478325 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.478337 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:11Z","lastTransitionTime":"2026-03-20T08:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.582132 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.582196 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.582218 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.582246 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.582269 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:11Z","lastTransitionTime":"2026-03-20T08:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.686138 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.686241 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.686262 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.686299 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.686333 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:11Z","lastTransitionTime":"2026-03-20T08:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.790018 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.790086 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.790100 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.790123 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.790138 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:11Z","lastTransitionTime":"2026-03-20T08:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.893024 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.893078 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.893099 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.893125 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.893142 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:11Z","lastTransitionTime":"2026-03-20T08:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.996134 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.996184 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.996197 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.996213 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:11 crc kubenswrapper[4823]: I0320 08:59:11.996227 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:11Z","lastTransitionTime":"2026-03-20T08:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.099826 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.099869 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.099890 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.099908 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.099921 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:12Z","lastTransitionTime":"2026-03-20T08:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.203609 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.203997 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.204142 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.204279 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.204396 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:12Z","lastTransitionTime":"2026-03-20T08:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.308335 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.308847 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.308973 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.309086 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.309201 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:12Z","lastTransitionTime":"2026-03-20T08:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.412676 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.412799 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.412820 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.412845 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.412859 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:12Z","lastTransitionTime":"2026-03-20T08:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.515592 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.515667 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.515688 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.515715 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.515731 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:12Z","lastTransitionTime":"2026-03-20T08:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.618651 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.618697 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.618708 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.618758 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.618770 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:12Z","lastTransitionTime":"2026-03-20T08:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.721501 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.721562 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.721580 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.721605 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.721626 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:12Z","lastTransitionTime":"2026-03-20T08:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.825638 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.825703 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.825720 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.825772 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.825790 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:12Z","lastTransitionTime":"2026-03-20T08:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.928860 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.928899 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.928912 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.928930 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:12 crc kubenswrapper[4823]: I0320 08:59:12.928939 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:12Z","lastTransitionTime":"2026-03-20T08:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.032011 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.032500 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.032685 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.032901 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.033112 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:13Z","lastTransitionTime":"2026-03-20T08:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.096525 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.096650 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:13 crc kubenswrapper[4823]: E0320 08:59:13.096712 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:13 crc kubenswrapper[4823]: E0320 08:59:13.096949 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.097005 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.097053 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:13 crc kubenswrapper[4823]: E0320 08:59:13.097135 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:13 crc kubenswrapper[4823]: E0320 08:59:13.097221 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.136867 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.136970 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.136992 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.137017 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.137035 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:13Z","lastTransitionTime":"2026-03-20T08:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.240307 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.240787 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.240933 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.241063 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.241195 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:13Z","lastTransitionTime":"2026-03-20T08:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.344292 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.344377 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.344388 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.344409 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.344422 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:13Z","lastTransitionTime":"2026-03-20T08:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.447276 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.447319 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.447331 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.447350 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.447368 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:13Z","lastTransitionTime":"2026-03-20T08:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.550467 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.550506 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.550517 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.550532 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.550544 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:13Z","lastTransitionTime":"2026-03-20T08:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.653462 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.653536 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.653551 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.653578 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.653594 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:13Z","lastTransitionTime":"2026-03-20T08:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.756235 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.756288 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.756298 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.756315 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.756325 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:13Z","lastTransitionTime":"2026-03-20T08:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.860090 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.860165 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.860180 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.860202 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.860219 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:13Z","lastTransitionTime":"2026-03-20T08:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.964350 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.964421 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.964441 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.964467 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:13 crc kubenswrapper[4823]: I0320 08:59:13.964485 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:13Z","lastTransitionTime":"2026-03-20T08:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:14 crc kubenswrapper[4823]: E0320 08:59:14.065074 4823 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.112041 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.126849 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.145239 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.170815 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:01Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.680919 6842 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:59:01.680927 6842 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:59:01.680932 6842 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:59:01.680939 6842 handler.go:208] Removed *v1.Node event handler 7\\\\nI0320 08:59:01.681072 6842 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0320 08:59:01.681284 6842 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681710 6842 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681769 6842 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.681789 6842 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.189105 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: E0320 08:59:14.197073 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.204334 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.218865 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.239240 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.263561 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.278877 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.295876 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.311664 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.326553 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.344493 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.357310 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.368497 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.377630 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.388995 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.904580 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.904617 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.904629 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.904644 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.904654 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:14Z","lastTransitionTime":"2026-03-20T08:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:14 crc kubenswrapper[4823]: E0320 08:59:14.918479 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.923465 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.923554 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.923571 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.923596 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.923614 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:14Z","lastTransitionTime":"2026-03-20T08:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:14 crc kubenswrapper[4823]: E0320 08:59:14.938275 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.942946 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.942985 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.943000 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.943020 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.943042 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:14Z","lastTransitionTime":"2026-03-20T08:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:14 crc kubenswrapper[4823]: E0320 08:59:14.957593 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.962179 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.962219 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.962232 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.962251 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.962262 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:14Z","lastTransitionTime":"2026-03-20T08:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:14 crc kubenswrapper[4823]: E0320 08:59:14.975154 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.978826 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.978881 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.978894 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.978915 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:14 crc kubenswrapper[4823]: I0320 08:59:14.978927 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:14Z","lastTransitionTime":"2026-03-20T08:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:14 crc kubenswrapper[4823]: E0320 08:59:14.991260 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:14Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:14 crc kubenswrapper[4823]: E0320 08:59:14.991450 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:59:15 crc kubenswrapper[4823]: I0320 08:59:15.096977 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:15 crc kubenswrapper[4823]: E0320 08:59:15.097103 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:15 crc kubenswrapper[4823]: I0320 08:59:15.096981 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:15 crc kubenswrapper[4823]: I0320 08:59:15.096982 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:15 crc kubenswrapper[4823]: E0320 08:59:15.097164 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:15 crc kubenswrapper[4823]: I0320 08:59:15.096997 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:15 crc kubenswrapper[4823]: E0320 08:59:15.097287 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:15 crc kubenswrapper[4823]: E0320 08:59:15.097314 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:17 crc kubenswrapper[4823]: I0320 08:59:17.097025 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:17 crc kubenswrapper[4823]: I0320 08:59:17.097113 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:17 crc kubenswrapper[4823]: E0320 08:59:17.097219 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:17 crc kubenswrapper[4823]: I0320 08:59:17.097255 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:17 crc kubenswrapper[4823]: I0320 08:59:17.097294 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:17 crc kubenswrapper[4823]: E0320 08:59:17.097475 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:17 crc kubenswrapper[4823]: E0320 08:59:17.097676 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:17 crc kubenswrapper[4823]: E0320 08:59:17.097849 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.098606 4823 scope.go:117] "RemoveContainer" containerID="916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.730579 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/1.log" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.733113 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041"} Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.733623 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.748124 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.758706 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.777021 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.789670 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.804777 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.819553 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.832878 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.843037 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.864155 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.882334 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.893306 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.901501 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.919163 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:01Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.680919 6842 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:59:01.680927 6842 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:59:01.680932 6842 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:59:01.680939 6842 handler.go:208] Removed *v1.Node event handler 7\\\\nI0320 08:59:01.681072 6842 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0320 08:59:01.681284 6842 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681710 6842 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681769 6842 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.681789 6842 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.928823 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.940706 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.950040 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.962886 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:18 crc kubenswrapper[4823]: I0320 08:59:18.974492 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:18Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.096918 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:19 crc kubenswrapper[4823]: E0320 08:59:19.097042 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.097214 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:19 crc kubenswrapper[4823]: E0320 08:59:19.097257 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.097357 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:19 crc kubenswrapper[4823]: E0320 08:59:19.097407 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.097502 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:19 crc kubenswrapper[4823]: E0320 08:59:19.097544 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.136990 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.152448 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.169259 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.187411 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: E0320 08:59:19.198416 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.199816 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.214202 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.235517 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.255304 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.274367 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.291651 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.310596 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.332336 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.348736 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.368789 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:01Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.680919 6842 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:59:01.680927 6842 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:59:01.680932 6842 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:59:01.680939 6842 handler.go:208] Removed *v1.Node event handler 7\\\\nI0320 08:59:01.681072 6842 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0320 08:59:01.681284 6842 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681710 6842 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681769 6842 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.681789 6842 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.382325 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.405023 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.421514 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.436153 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.446570 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.740737 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/2.log" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.742245 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/1.log" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.748410 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041" exitCode=1 Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.748455 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041"} Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.748489 4823 scope.go:117] "RemoveContainer" containerID="916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.749315 4823 scope.go:117] "RemoveContainer" containerID="bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041" Mar 20 08:59:19 crc kubenswrapper[4823]: E0320 08:59:19.749504 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.765047 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.778417 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.796312 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.807987 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.818092 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.830552 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.839969 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.849168 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.860486 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.872570 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.887279 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.899289 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.916481 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://916f71c46e9a219c65121cbf4fcbf4d3730eb642ae53e84d69607844242c2ecf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:01Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.680919 6842 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:59:01.680927 6842 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:59:01.680932 6842 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:59:01.680939 6842 handler.go:208] Removed *v1.Node event handler 7\\\\nI0320 08:59:01.681072 6842 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0320 08:59:01.681284 6842 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681710 6842 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:01.681769 6842 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:01.681789 6842 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"message\\\":\\\"ry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066658 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066662 7043 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-fb4mg in node crc\\\\nI0320 08:59:19.066665 7043 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0320 08:59:19.066674 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066681 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066666 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg after 0 failed attempt(s)\\\\nI0320 08:59:19.066688 7043 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-7wn7x in node crc\\\\nI0320 08:59:19.066693 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7wn7x after 0 failed attempt(s)\\\\nI0320 08:59:19.066698 7043 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066705 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.926537 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.946764 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.959856 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.975038 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:19 crc kubenswrapper[4823]: I0320 08:59:19.988839 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.756369 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/2.log" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.761841 4823 scope.go:117] "RemoveContainer" containerID="bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041" Mar 20 08:59:20 crc kubenswrapper[4823]: E0320 08:59:20.762255 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.783473 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.819802 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"message\\\":\\\"ry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066658 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066662 7043 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-fb4mg in node crc\\\\nI0320 08:59:19.066665 7043 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0320 08:59:19.066674 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066681 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066666 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg after 0 failed attempt(s)\\\\nI0320 08:59:19.066688 7043 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-7wn7x in node crc\\\\nI0320 08:59:19.066693 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7wn7x after 0 failed attempt(s)\\\\nI0320 08:59:19.066698 7043 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066705 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.836658 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.855448 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.873293 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.891069 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.910448 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.926055 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.941930 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.973019 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4823]: I0320 08:59:20.988567 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.000114 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.015328 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.034074 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.047592 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.060010 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.076049 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.089481 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.096801 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.096891 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.096910 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.096968 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:21 crc kubenswrapper[4823]: E0320 08:59:21.097128 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:21 crc kubenswrapper[4823]: E0320 08:59:21.097387 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:21 crc kubenswrapper[4823]: E0320 08:59:21.097671 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:21 crc kubenswrapper[4823]: E0320 08:59:21.097727 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:21 crc kubenswrapper[4823]: I0320 08:59:21.109918 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 20 08:59:23 crc kubenswrapper[4823]: I0320 08:59:23.097164 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:23 crc kubenswrapper[4823]: I0320 08:59:23.097204 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:23 crc kubenswrapper[4823]: I0320 08:59:23.097249 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:23 crc kubenswrapper[4823]: I0320 08:59:23.097217 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:23 crc kubenswrapper[4823]: E0320 08:59:23.097374 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:23 crc kubenswrapper[4823]: E0320 08:59:23.097471 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:23 crc kubenswrapper[4823]: E0320 08:59:23.097574 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:23 crc kubenswrapper[4823]: E0320 08:59:23.097700 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.135137 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.156708 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.183022 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: E0320 08:59:24.199819 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.211385 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"message\\\":\\\"ry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066658 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066662 7043 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-fb4mg in node crc\\\\nI0320 08:59:19.066665 7043 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0320 08:59:19.066674 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066681 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066666 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg after 0 failed attempt(s)\\\\nI0320 08:59:19.066688 7043 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-7wn7x in node crc\\\\nI0320 08:59:19.066693 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7wn7x after 0 failed attempt(s)\\\\nI0320 08:59:19.066698 7043 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066705 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.223360 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.237727 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.250659 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.263032 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.282669 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.294345 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.307547 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.321241 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49a4ae2-55cd-4af1-a557-a29d4b125c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e8552c1d3d581d0f496b5f2cc3da10b7112c980788babf34c126bc9d3bc624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca0df1d0295ef37be29121ac847de5209abb269b53439eb479ffb9b82dad1a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1bdf60e1820b32bf856c731144b58c5a133edb212cd21952f53c679eef58699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.337704 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.349821 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.359608 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.372401 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.384197 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.396621 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4823]: I0320 08:59:24.408457 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.096973 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.097030 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.097054 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.097145 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:25 crc kubenswrapper[4823]: E0320 08:59:25.097779 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:25 crc kubenswrapper[4823]: E0320 08:59:25.097971 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:25 crc kubenswrapper[4823]: E0320 08:59:25.098154 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:25 crc kubenswrapper[4823]: E0320 08:59:25.098289 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.376880 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.376937 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.376960 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.376986 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.377007 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:25Z","lastTransitionTime":"2026-03-20T08:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:25 crc kubenswrapper[4823]: E0320 08:59:25.398163 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.403850 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.403983 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.404004 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.404027 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.404042 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:25Z","lastTransitionTime":"2026-03-20T08:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:25 crc kubenswrapper[4823]: E0320 08:59:25.418954 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.422157 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.422194 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.422205 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.422221 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.422267 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:25Z","lastTransitionTime":"2026-03-20T08:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:25 crc kubenswrapper[4823]: E0320 08:59:25.435181 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.438491 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.438533 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.438545 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.438562 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.438573 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:25Z","lastTransitionTime":"2026-03-20T08:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:25 crc kubenswrapper[4823]: E0320 08:59:25.450004 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.453525 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.453553 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.453564 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.453579 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:25 crc kubenswrapper[4823]: I0320 08:59:25.453590 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:25Z","lastTransitionTime":"2026-03-20T08:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:25 crc kubenswrapper[4823]: E0320 08:59:25.471827 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4823]: E0320 08:59:25.472014 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:59:27 crc kubenswrapper[4823]: I0320 08:59:27.096979 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:27 crc kubenswrapper[4823]: I0320 08:59:27.097057 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:27 crc kubenswrapper[4823]: I0320 08:59:27.096988 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:27 crc kubenswrapper[4823]: I0320 08:59:27.096980 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:27 crc kubenswrapper[4823]: E0320 08:59:27.097212 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:27 crc kubenswrapper[4823]: E0320 08:59:27.097336 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:27 crc kubenswrapper[4823]: E0320 08:59:27.097504 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:27 crc kubenswrapper[4823]: E0320 08:59:27.097658 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:29 crc kubenswrapper[4823]: I0320 08:59:29.096528 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:29 crc kubenswrapper[4823]: I0320 08:59:29.096617 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:29 crc kubenswrapper[4823]: I0320 08:59:29.096691 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:29 crc kubenswrapper[4823]: E0320 08:59:29.096858 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:29 crc kubenswrapper[4823]: I0320 08:59:29.096921 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:29 crc kubenswrapper[4823]: E0320 08:59:29.097047 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:29 crc kubenswrapper[4823]: E0320 08:59:29.097189 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:29 crc kubenswrapper[4823]: E0320 08:59:29.097294 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:29 crc kubenswrapper[4823]: E0320 08:59:29.201816 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:31 crc kubenswrapper[4823]: I0320 08:59:31.096452 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:31 crc kubenswrapper[4823]: I0320 08:59:31.096477 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:31 crc kubenswrapper[4823]: E0320 08:59:31.096560 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:31 crc kubenswrapper[4823]: I0320 08:59:31.096455 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:31 crc kubenswrapper[4823]: E0320 08:59:31.096668 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:31 crc kubenswrapper[4823]: I0320 08:59:31.096691 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:31 crc kubenswrapper[4823]: E0320 08:59:31.096717 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:31 crc kubenswrapper[4823]: E0320 08:59:31.096775 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:33 crc kubenswrapper[4823]: I0320 08:59:33.096309 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:33 crc kubenswrapper[4823]: I0320 08:59:33.096372 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:33 crc kubenswrapper[4823]: I0320 08:59:33.096379 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:33 crc kubenswrapper[4823]: E0320 08:59:33.096576 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:33 crc kubenswrapper[4823]: I0320 08:59:33.096606 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:33 crc kubenswrapper[4823]: E0320 08:59:33.096697 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:33 crc kubenswrapper[4823]: E0320 08:59:33.096969 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:33 crc kubenswrapper[4823]: E0320 08:59:33.097083 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.117649 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.138474 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"message\\\":\\\"ry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066658 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066662 7043 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-fb4mg in node crc\\\\nI0320 08:59:19.066665 7043 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0320 08:59:19.066674 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066681 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066666 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg after 0 failed attempt(s)\\\\nI0320 08:59:19.066688 7043 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-7wn7x in node crc\\\\nI0320 08:59:19.066693 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7wn7x after 0 failed attempt(s)\\\\nI0320 08:59:19.066698 7043 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066705 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.156239 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.182349 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.197066 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: E0320 08:59:34.202651 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.216496 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.235337 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.250841 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.273296 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.288294 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49a4ae2-55cd-4af1-a557-a29d4b125c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e8552c1d3d581d0f496b5f2cc3da10b7112c980788babf34c126bc9d3bc624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca0df1d0295ef37be29121ac847de5209abb269b53439eb479ffb9b82dad1a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1bdf60e1820b32bf856c731144b58c5a133edb212cd21952f53c679eef58699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.314365 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.368154 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.391593 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.408705 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.422194 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.435813 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.451609 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.466009 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:34 crc kubenswrapper[4823]: I0320 08:59:34.479712 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:34Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.096774 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.096827 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.096826 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.096801 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:35 crc kubenswrapper[4823]: E0320 08:59:35.096918 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:35 crc kubenswrapper[4823]: E0320 08:59:35.097011 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:35 crc kubenswrapper[4823]: E0320 08:59:35.097070 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:35 crc kubenswrapper[4823]: E0320 08:59:35.097146 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.538255 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.539643 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.539849 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.540005 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.540142 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:35Z","lastTransitionTime":"2026-03-20T08:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:35 crc kubenswrapper[4823]: E0320 08:59:35.561081 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:35Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.567049 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.567106 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.567128 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.567157 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.567178 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:35Z","lastTransitionTime":"2026-03-20T08:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:35 crc kubenswrapper[4823]: E0320 08:59:35.590788 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:35Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.596426 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.596493 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.596512 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.596538 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.596555 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:35Z","lastTransitionTime":"2026-03-20T08:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:35 crc kubenswrapper[4823]: E0320 08:59:35.618232 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:35Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.623985 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.624224 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.624257 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.624294 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.624318 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:35Z","lastTransitionTime":"2026-03-20T08:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:35 crc kubenswrapper[4823]: E0320 08:59:35.646472 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:35Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.651897 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.651958 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.651972 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.651996 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:35 crc kubenswrapper[4823]: I0320 08:59:35.652011 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:35Z","lastTransitionTime":"2026-03-20T08:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:35 crc kubenswrapper[4823]: E0320 08:59:35.669107 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:35Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:35 crc kubenswrapper[4823]: E0320 08:59:35.669327 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:59:36 crc kubenswrapper[4823]: I0320 08:59:36.098640 4823 scope.go:117] "RemoveContainer" containerID="bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041" Mar 20 08:59:36 crc kubenswrapper[4823]: E0320 08:59:36.099531 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 08:59:37 crc kubenswrapper[4823]: I0320 08:59:37.096291 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:37 crc kubenswrapper[4823]: I0320 08:59:37.096360 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:37 crc kubenswrapper[4823]: I0320 08:59:37.096508 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:37 crc kubenswrapper[4823]: E0320 08:59:37.096831 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:37 crc kubenswrapper[4823]: E0320 08:59:37.097032 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:37 crc kubenswrapper[4823]: E0320 08:59:37.097195 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:37 crc kubenswrapper[4823]: I0320 08:59:37.097918 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:37 crc kubenswrapper[4823]: E0320 08:59:37.098639 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:39 crc kubenswrapper[4823]: I0320 08:59:39.096617 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:39 crc kubenswrapper[4823]: I0320 08:59:39.096680 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:39 crc kubenswrapper[4823]: I0320 08:59:39.096637 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:39 crc kubenswrapper[4823]: E0320 08:59:39.097035 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:39 crc kubenswrapper[4823]: E0320 08:59:39.096872 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:39 crc kubenswrapper[4823]: I0320 08:59:39.097128 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:39 crc kubenswrapper[4823]: E0320 08:59:39.097251 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:39 crc kubenswrapper[4823]: E0320 08:59:39.097364 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:39 crc kubenswrapper[4823]: E0320 08:59:39.204327 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.835861 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9vsj7_d33f9d63-34d7-4722-95d5-13bdcb32b99a/kube-multus/0.log" Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.835945 4823 generic.go:334] "Generic (PLEG): container finished" podID="d33f9d63-34d7-4722-95d5-13bdcb32b99a" containerID="05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8" exitCode=1 Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.835990 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9vsj7" event={"ID":"d33f9d63-34d7-4722-95d5-13bdcb32b99a","Type":"ContainerDied","Data":"05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8"} Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.836883 4823 scope.go:117] "RemoveContainer" containerID="05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8" Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.860953 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.894882 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"message\\\":\\\"ry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066658 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066662 7043 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-fb4mg in node crc\\\\nI0320 08:59:19.066665 7043 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0320 08:59:19.066674 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066681 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066666 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg after 0 failed attempt(s)\\\\nI0320 08:59:19.066688 7043 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-7wn7x in node crc\\\\nI0320 08:59:19.066693 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7wn7x after 0 failed attempt(s)\\\\nI0320 08:59:19.066698 7043 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066705 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.909216 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.924825 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.934885 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.948116 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.964181 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.976414 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4823]: I0320 08:59:40.997810 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.012698 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49a4ae2-55cd-4af1-a557-a29d4b125c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e8552c1d3d581d0f496b5f2cc3da10b7112c980788babf34c126bc9d3bc624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca0df1d0295ef37be29121ac847de5209abb269b53439eb479ffb9b82dad1a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1bdf60e1820b32bf856c731144b58c5a133edb212cd21952f53c679eef58699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.031012 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.044452 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.055242 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.068022 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7\\\\n2026-03-20T08:58:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.080897 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.092220 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.100641 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:41 crc kubenswrapper[4823]: E0320 08:59:41.100860 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.101021 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:41 crc kubenswrapper[4823]: E0320 08:59:41.101137 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.101308 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.101370 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:41 crc kubenswrapper[4823]: E0320 08:59:41.101954 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:41 crc kubenswrapper[4823]: E0320 08:59:41.102036 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.108484 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.124457 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.137115 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.860897 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9vsj7_d33f9d63-34d7-4722-95d5-13bdcb32b99a/kube-multus/0.log" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.861017 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9vsj7" event={"ID":"d33f9d63-34d7-4722-95d5-13bdcb32b99a","Type":"ContainerStarted","Data":"78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b"} Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.886067 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.903810 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.919487 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49a4ae2-55cd-4af1-a557-a29d4b125c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e8552c1d3d581d0f496b5f2cc3da10b7112c980788babf34c126bc9d3bc624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca0df1d0295ef37be29121ac847de5209abb269b53439eb479ffb9b82dad1a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1bdf60e1820b32bf856c731144b58c5a133edb212cd21952f53c679eef58699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.951886 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.966857 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4823]: I0320 08:59:41.987865 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.005323 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.020021 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.036565 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.050608 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.066411 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.083379 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.098049 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7\\\\n2026-03-20T08:58:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.113183 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.132845 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.144814 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.160388 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.176549 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.206372 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"message\\\":\\\"ry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066658 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066662 7043 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-fb4mg in node crc\\\\nI0320 08:59:19.066665 7043 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0320 08:59:19.066674 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066681 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066666 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg after 0 failed attempt(s)\\\\nI0320 08:59:19.066688 7043 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-7wn7x in node crc\\\\nI0320 08:59:19.066693 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7wn7x after 0 failed attempt(s)\\\\nI0320 08:59:19.066698 7043 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066705 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:42 crc kubenswrapper[4823]: I0320 08:59:42.975670 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:42 crc kubenswrapper[4823]: E0320 08:59:42.975837 4823 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:59:42 crc kubenswrapper[4823]: E0320 08:59:42.975941 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 09:00:46.975923204 +0000 UTC m=+213.661819578 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:59:43 crc kubenswrapper[4823]: I0320 08:59:43.077330 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:59:43 crc kubenswrapper[4823]: I0320 08:59:43.077470 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.077558 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:47.077516777 +0000 UTC m=+213.763413191 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.077678 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.077705 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:59:43 crc kubenswrapper[4823]: I0320 08:59:43.077702 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:43 crc kubenswrapper[4823]: I0320 08:59:43.077805 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.077724 4823 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.077977 4823 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.077867 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.078070 4823 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.078091 4823 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.078064 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 09:00:47.078017343 +0000 UTC m=+213.763913747 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.078207 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 09:00:47.078165988 +0000 UTC m=+213.764062402 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.078254 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 09:00:47.0782353 +0000 UTC m=+213.764131824 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:43 crc kubenswrapper[4823]: I0320 08:59:43.096641 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:43 crc kubenswrapper[4823]: I0320 08:59:43.096643 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:43 crc kubenswrapper[4823]: I0320 08:59:43.096725 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:43 crc kubenswrapper[4823]: I0320 08:59:43.096846 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.097006 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.097191 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.097304 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.097468 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:43 crc kubenswrapper[4823]: I0320 08:59:43.178827 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.179069 4823 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:59:43 crc kubenswrapper[4823]: E0320 08:59:43.179214 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs podName:316f49ed-2e3c-42b7-a133-a36c3074810e nodeName:}" failed. No retries permitted until 2026-03-20 09:00:47.179186222 +0000 UTC m=+213.865082636 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs") pod "network-metrics-daemon-jzzk2" (UID: "316f49ed-2e3c-42b7-a133-a36c3074810e") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.109100 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.127866 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.146243 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.163570 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.180799 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7\\\\n2026-03-20T08:58:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.195638 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: E0320 08:59:44.205375 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.213329 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.228866 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.240497 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.256886 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.273704 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.301322 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"message\\\":\\\"ry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066658 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066662 7043 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-fb4mg in node crc\\\\nI0320 08:59:19.066665 7043 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0320 08:59:19.066674 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066681 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066666 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg after 0 failed attempt(s)\\\\nI0320 08:59:19.066688 7043 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-7wn7x in node crc\\\\nI0320 08:59:19.066693 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7wn7x after 0 failed attempt(s)\\\\nI0320 08:59:19.066698 7043 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066705 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.313056 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.332124 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.345792 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.358361 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49a4ae2-55cd-4af1-a557-a29d4b125c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e8552c1d3d581d0f496b5f2cc3da10b7112c980788babf34c126bc9d3bc624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca0df1d0295ef37be29121ac847de5209abb269b53439eb479ffb9b82dad1a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1bdf60e1820b32bf856c731144b58c5a133edb212cd21952f53c679eef58699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.380185 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.394656 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:44 crc kubenswrapper[4823]: I0320 08:59:44.416275 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:45 crc kubenswrapper[4823]: I0320 08:59:45.096657 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:45 crc kubenswrapper[4823]: I0320 08:59:45.096717 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:45 crc kubenswrapper[4823]: I0320 08:59:45.096810 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:45 crc kubenswrapper[4823]: I0320 08:59:45.096852 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:45 crc kubenswrapper[4823]: E0320 08:59:45.097075 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:45 crc kubenswrapper[4823]: E0320 08:59:45.097173 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:45 crc kubenswrapper[4823]: E0320 08:59:45.097329 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:45 crc kubenswrapper[4823]: E0320 08:59:45.097438 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.055011 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.055084 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.055121 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.055146 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.055165 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:46Z","lastTransitionTime":"2026-03-20T08:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:46 crc kubenswrapper[4823]: E0320 08:59:46.078570 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.084149 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.084195 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.084210 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.084230 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.084247 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:46Z","lastTransitionTime":"2026-03-20T08:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:46 crc kubenswrapper[4823]: E0320 08:59:46.104658 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.110052 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.110098 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.110108 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.110129 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.110140 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:46Z","lastTransitionTime":"2026-03-20T08:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:46 crc kubenswrapper[4823]: E0320 08:59:46.131872 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.136807 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.136881 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.136908 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.136943 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.136964 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:46Z","lastTransitionTime":"2026-03-20T08:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:46 crc kubenswrapper[4823]: E0320 08:59:46.156338 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.160287 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.160319 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.160330 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.160347 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:46 crc kubenswrapper[4823]: I0320 08:59:46.160359 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:46Z","lastTransitionTime":"2026-03-20T08:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:46 crc kubenswrapper[4823]: E0320 08:59:46.180034 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:46 crc kubenswrapper[4823]: E0320 08:59:46.180175 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:59:47 crc kubenswrapper[4823]: I0320 08:59:47.096418 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:47 crc kubenswrapper[4823]: I0320 08:59:47.096467 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:47 crc kubenswrapper[4823]: I0320 08:59:47.096428 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:47 crc kubenswrapper[4823]: I0320 08:59:47.096428 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:47 crc kubenswrapper[4823]: E0320 08:59:47.096619 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:47 crc kubenswrapper[4823]: E0320 08:59:47.096810 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:47 crc kubenswrapper[4823]: E0320 08:59:47.096969 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:47 crc kubenswrapper[4823]: E0320 08:59:47.097120 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:48 crc kubenswrapper[4823]: I0320 08:59:48.100321 4823 scope.go:117] "RemoveContainer" containerID="bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041" Mar 20 08:59:48 crc kubenswrapper[4823]: I0320 08:59:48.894553 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/2.log" Mar 20 08:59:48 crc kubenswrapper[4823]: I0320 08:59:48.897890 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469"} Mar 20 08:59:48 crc kubenswrapper[4823]: I0320 08:59:48.898334 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 08:59:48 crc kubenswrapper[4823]: I0320 08:59:48.916370 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:48 crc kubenswrapper[4823]: I0320 08:59:48.931790 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:48 crc kubenswrapper[4823]: I0320 08:59:48.949305 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49a4ae2-55cd-4af1-a557-a29d4b125c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e8552c1d3d581d0f496b5f2cc3da10b7112c980788babf34c126bc9d3bc624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca0df1d0295ef37be29121ac847de5209abb269b53439eb479ffb9b82dad1a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1bdf60e1820b32bf856c731144b58c5a133edb212cd21952f53c679eef58699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:48 crc kubenswrapper[4823]: I0320 08:59:48.972345 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:48 crc kubenswrapper[4823]: I0320 08:59:48.984677 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:48 crc kubenswrapper[4823]: I0320 08:59:48.999837 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.014687 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7\\\\n2026-03-20T08:58:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.028104 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.040014 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.053026 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.065495 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.080459 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.096774 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.096790 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.096794 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.096831 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:49 crc kubenswrapper[4823]: E0320 08:59:49.097397 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:49 crc kubenswrapper[4823]: E0320 08:59:49.097461 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:49 crc kubenswrapper[4823]: E0320 08:59:49.097650 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:49 crc kubenswrapper[4823]: E0320 08:59:49.097724 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.099099 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.126150 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"message\\\":\\\"ry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066658 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066662 7043 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-fb4mg in node crc\\\\nI0320 08:59:19.066665 7043 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0320 08:59:19.066674 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066681 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066666 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg after 0 failed attempt(s)\\\\nI0320 08:59:19.066688 7043 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-7wn7x in node crc\\\\nI0320 08:59:19.066693 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7wn7x after 0 failed attempt(s)\\\\nI0320 08:59:19.066698 7043 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066705 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.140099 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.156365 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.170125 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.183665 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.196124 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: E0320 08:59:49.206986 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.906841 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/3.log" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.908187 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/2.log" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.913192 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" exitCode=1 Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.913264 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469"} Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.913343 4823 scope.go:117] "RemoveContainer" containerID="bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.914584 4823 scope.go:117] "RemoveContainer" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 08:59:49 crc kubenswrapper[4823]: E0320 08:59:49.914950 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.929379 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49a4ae2-55cd-4af1-a557-a29d4b125c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e8552c1d3d581d0f496b5f2cc3da10b7112c980788babf34c126bc9d3bc624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca0df1d0295ef37be29121ac847de5209abb269b53439eb479ffb9b82dad1a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1bdf60e1820b32bf856c731144b58c5a133edb212cd21952f53c679eef58699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.962050 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.979964 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:49 crc kubenswrapper[4823]: I0320 08:59:49.995043 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.012182 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7\\\\n2026-03-20T08:58:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.025133 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.038368 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.053539 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.074118 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.090197 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.101377 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.122518 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb77d76b6bd583d54bafdadad3e037449275329e61997fb351587cfb03f3c041\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"message\\\":\\\"ry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066658 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066662 7043 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-fb4mg in node crc\\\\nI0320 08:59:19.066665 7043 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0320 08:59:19.066674 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066681 7043 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-7wn7x\\\\nI0320 08:59:19.066666 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-fb4mg after 0 failed attempt(s)\\\\nI0320 08:59:19.066688 7043 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-7wn7x in node crc\\\\nI0320 08:59:19.066693 7043 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-7wn7x after 0 failed attempt(s)\\\\nI0320 08:59:19.066698 7043 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-fb4mg\\\\nI0320 08:59:19.066705 7043 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:49Z\\\",\\\"message\\\":\\\" time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z]\\\\nI0320 08:59:49.108935 7367 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"ebd4748e-0473-49fb-88ad-83dbb221791a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.88\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 08:59:49.10897\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.138938 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.155936 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.170563 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.186158 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.203790 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.219492 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.234035 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.919565 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/3.log" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.923822 4823 scope.go:117] "RemoveContainer" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 08:59:50 crc kubenswrapper[4823]: E0320 08:59:50.924038 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.940231 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.957590 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:50 crc kubenswrapper[4823]: I0320 08:59:50.977247 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49a4ae2-55cd-4af1-a557-a29d4b125c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e8552c1d3d581d0f496b5f2cc3da10b7112c980788babf34c126bc9d3bc624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca0df1d0295ef37be29121ac847de5209abb269b53439eb479ffb9b82dad1a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1bdf60e1820b32bf856c731144b58c5a133edb212cd21952f53c679eef58699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.002315 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.021285 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.041267 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.060072 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7\\\\n2026-03-20T08:58:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.077068 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.093310 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.096282 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.096294 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.096333 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:51 crc kubenswrapper[4823]: E0320 08:59:51.096408 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.096427 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:51 crc kubenswrapper[4823]: E0320 08:59:51.096592 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:51 crc kubenswrapper[4823]: E0320 08:59:51.096881 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:51 crc kubenswrapper[4823]: E0320 08:59:51.096958 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.112130 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.132843 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.151639 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.165411 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.188840 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:49Z\\\",\\\"message\\\":\\\" time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z]\\\\nI0320 08:59:49.108935 7367 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"ebd4748e-0473-49fb-88ad-83dbb221791a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.88\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 08:59:49.10897\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.205603 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.218975 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.231914 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.253265 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:51 crc kubenswrapper[4823]: I0320 08:59:51.273986 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:53 crc kubenswrapper[4823]: I0320 08:59:53.097002 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:53 crc kubenswrapper[4823]: I0320 08:59:53.097008 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:53 crc kubenswrapper[4823]: E0320 08:59:53.097103 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:53 crc kubenswrapper[4823]: I0320 08:59:53.097149 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:53 crc kubenswrapper[4823]: E0320 08:59:53.097172 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:53 crc kubenswrapper[4823]: I0320 08:59:53.097022 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:53 crc kubenswrapper[4823]: E0320 08:59:53.097296 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:53 crc kubenswrapper[4823]: E0320 08:59:53.097335 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.111825 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.121330 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.135613 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.147249 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.168060 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:49Z\\\",\\\"message\\\":\\\" time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z]\\\\nI0320 08:59:49.108935 7367 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"ebd4748e-0473-49fb-88ad-83dbb221791a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.88\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 08:59:49.10897\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.179817 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.196147 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.206877 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: E0320 08:59:54.208023 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.223188 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49a4ae2-55cd-4af1-a557-a29d4b125c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e8552c1d3d581d0f496b5f2cc3da10b7112c980788babf34c126bc9d3bc624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca0df1d0295ef37be29121ac847de5209abb269b53439eb479ffb9b82dad1a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1bdf60e1820b32bf856c731144b58c5a133edb212cd21952f53c679eef58699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.239708 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.249846 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.262594 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.273997 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.289041 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.300680 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.310582 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.322115 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7\\\\n2026-03-20T08:58:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.332020 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:54 crc kubenswrapper[4823]: I0320 08:59:54.343275 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:55 crc kubenswrapper[4823]: I0320 08:59:55.097013 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:55 crc kubenswrapper[4823]: I0320 08:59:55.097111 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:55 crc kubenswrapper[4823]: I0320 08:59:55.097209 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:55 crc kubenswrapper[4823]: E0320 08:59:55.097207 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:55 crc kubenswrapper[4823]: E0320 08:59:55.097345 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:55 crc kubenswrapper[4823]: I0320 08:59:55.097407 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:55 crc kubenswrapper[4823]: E0320 08:59:55.097485 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:55 crc kubenswrapper[4823]: E0320 08:59:55.097565 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.212337 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.212391 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.212403 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.212423 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.212437 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:56Z","lastTransitionTime":"2026-03-20T08:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:56 crc kubenswrapper[4823]: E0320 08:59:56.233009 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.239454 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.239539 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.239565 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.239597 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.239621 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:56Z","lastTransitionTime":"2026-03-20T08:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:56 crc kubenswrapper[4823]: E0320 08:59:56.261294 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.270904 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.270959 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.270971 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.270988 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.270999 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:56Z","lastTransitionTime":"2026-03-20T08:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:56 crc kubenswrapper[4823]: E0320 08:59:56.288682 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.293532 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.293590 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.293601 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.293621 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.293635 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:56Z","lastTransitionTime":"2026-03-20T08:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:56 crc kubenswrapper[4823]: E0320 08:59:56.314039 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.320496 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.320544 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.320558 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.320580 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:56 crc kubenswrapper[4823]: I0320 08:59:56.320595 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:56Z","lastTransitionTime":"2026-03-20T08:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:56 crc kubenswrapper[4823]: E0320 08:59:56.341631 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:56Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:56 crc kubenswrapper[4823]: E0320 08:59:56.342038 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:59:57 crc kubenswrapper[4823]: I0320 08:59:57.097022 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:57 crc kubenswrapper[4823]: E0320 08:59:57.097516 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:57 crc kubenswrapper[4823]: I0320 08:59:57.097092 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:57 crc kubenswrapper[4823]: E0320 08:59:57.097651 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:57 crc kubenswrapper[4823]: I0320 08:59:57.097123 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:57 crc kubenswrapper[4823]: I0320 08:59:57.097033 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:57 crc kubenswrapper[4823]: E0320 08:59:57.097827 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:57 crc kubenswrapper[4823]: E0320 08:59:57.097931 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:59 crc kubenswrapper[4823]: I0320 08:59:59.096589 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 08:59:59 crc kubenswrapper[4823]: I0320 08:59:59.096675 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:59 crc kubenswrapper[4823]: I0320 08:59:59.096816 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:59 crc kubenswrapper[4823]: E0320 08:59:59.096864 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 08:59:59 crc kubenswrapper[4823]: I0320 08:59:59.096915 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:59 crc kubenswrapper[4823]: E0320 08:59:59.097033 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:59 crc kubenswrapper[4823]: E0320 08:59:59.097177 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:59 crc kubenswrapper[4823]: E0320 08:59:59.097574 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:59 crc kubenswrapper[4823]: E0320 08:59:59.210581 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:01 crc kubenswrapper[4823]: I0320 09:00:01.096257 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:01 crc kubenswrapper[4823]: I0320 09:00:01.096260 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:01 crc kubenswrapper[4823]: I0320 09:00:01.096309 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:01 crc kubenswrapper[4823]: I0320 09:00:01.096577 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:01 crc kubenswrapper[4823]: E0320 09:00:01.096802 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:01 crc kubenswrapper[4823]: E0320 09:00:01.096835 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:01 crc kubenswrapper[4823]: E0320 09:00:01.096896 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:01 crc kubenswrapper[4823]: E0320 09:00:01.097069 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:03 crc kubenswrapper[4823]: I0320 09:00:03.097312 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:03 crc kubenswrapper[4823]: I0320 09:00:03.097359 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:03 crc kubenswrapper[4823]: I0320 09:00:03.097384 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:03 crc kubenswrapper[4823]: I0320 09:00:03.097312 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:03 crc kubenswrapper[4823]: E0320 09:00:03.097525 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:03 crc kubenswrapper[4823]: E0320 09:00:03.097651 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:03 crc kubenswrapper[4823]: E0320 09:00:03.097811 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:03 crc kubenswrapper[4823]: E0320 09:00:03.097875 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.099333 4823 scope.go:117] "RemoveContainer" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 09:00:04 crc kubenswrapper[4823]: E0320 09:00:04.099520 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.112866 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.136778 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bed1862-5bcd-46fe-bc84-2dbab416086b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:49Z\\\",\\\"message\\\":\\\" time 2026-03-20T08:59:49Z is after 2025-08-24T17:21:41Z]\\\\nI0320 08:59:49.108935 7367 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"ebd4748e-0473-49fb-88ad-83dbb221791a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.88\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 08:59:49.10897\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2x54k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.151245 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"316f49ed-2e3c-42b7-a133-a36c3074810e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p4wm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jzzk2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.169157 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebec5902-d4d5-4870-bfc3-031eb8d30ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a78a65ff2fe1bddead53a3c0f65f57261d754031929c1c7fd09454dfe1c0296\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e7865cbb6d04a273c0a9b535e74420844bf46fa205517474108d59cc07fe1dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:16.416903 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:16.418802 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:16.467295 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:16.472591 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0320 08:57:46.755360 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77a2a63e0b2c0dd16f5b645bc70c5edd44e6483a40cb1664326394fdd442182\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b9ce42c2db6a135d92028eef786682e8cb1761b3e48b40c2c839e7cbe50c519\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.185169 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d4698e7-b143-4cdf-a27a-0af19bbe6388\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0964eff7a5f9355cc5c52f326f6b01fdb45450233872821abb103b9f88a5dee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc5fa3a1d414bff246a112af0a66a73992739ca7e7425c108093e3247d3ef1a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.211066 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2d4857c2b0eb40a07ef3cbffd0f6d007169e4f4744bae5db3d66ede0387e662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: E0320 09:00:04.211788 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.228964 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d9b9db9-41b1-4510-bfe7-f57b07b66749\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:15Z\\\",\\\"message\\\":\\\"file observer\\\\nW0320 08:58:14.816679 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:58:14.816853 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:58:14.817783 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-198596482/tls.crt::/tmp/serving-cert-198596482/tls.key\\\\\\\"\\\\nI0320 08:58:15.075954 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:15.078343 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:15.078364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:15.078419 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:15.078437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:15.083541 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:15.083602 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:15.083629 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:15.083638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:15.083647 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:15.083657 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 08:58:15.083558 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 08:58:15.085463 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.243451 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74dc9b8c-5cfe-4336-863d-75c2a0ce028b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8283509f7f21294b5ab88fb9241c4feefd6c2a0339ce16baf42677c36cf2aab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c6d8fbe614b33b28ece151ee68215f6b0e1176db34395289fa3c64d9fda234e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dllqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cx7cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.259975 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38298bb6-9157-4c18-94a2-fedec7e95a4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f24d43a7b4348dc29357ae42fb27d9839814cabf5e9ab779e819f3b96792b174\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c069053d25425f8b628fef6f69e5a9a5d31b88facc53f9490846079b2890d808\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd5a4a6d87e43a6b6c68b7dd5329c879dc59aac4d185814273e7f7742ab7991e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27f940870425480737b8f730bde69922094b1c19c38d19136423c0a01a052766\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948f174aa415e29cf057094d09a6007cf6fb78f3d69aabab34c177e30aa6aa2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://056b3fc133d45efcf1a3f05de06ffe3121f3064c619b39c0f5345b18ec6d3659\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569a10dea525f0b2d60bfe08976976fc7af8d257a7c403aa6963e0367b20b24e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gfvd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fb4mg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.277068 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a49a4ae2-55cd-4af1-a557-a29d4b125c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61e8552c1d3d581d0f496b5f2cc3da10b7112c980788babf34c126bc9d3bc624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca0df1d0295ef37be29121ac847de5209abb269b53439eb479ffb9b82dad1a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1bdf60e1820b32bf856c731144b58c5a133edb212cd21952f53c679eef58699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5db3069c0463cbad51afcb05f0838474004d3f1108e557935c46cf98deacb38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.301681 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"846029e1-3b05-483a-bb9e-8c7badd85e12\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d13cde6414c74eb647f1b268a8f16bd699495214ca1947f5cde73d2197782c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0d089b3e5c37a11a0cd9e876625414a7e4a27fc4c64e4965fc36d89b5db7fbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55c534a8d3aa79b96077f598e7df4eae6e1a5ba7d59a5a916ce6d209b1fa10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f308e23b2a78d0d35814060c4f0a58d00b7a617d0508b6dff046a2126dff1ba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://60cda221b2d741d15792e8f8305962b9531566f58eb1a885189fff3a8e87510d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c5803200c2119bcd420ec809454894f634f0bfc8ac823dec786651dca1e9d1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35addd4270750e220897f94b880dc6cce0be089b61cccd3c2b11e317b11044f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:16Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0303530193cad1605b7dd0c32630bb8bb3052803a7a6a803876da1fd17c688e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.326236 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.341070 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wn7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c32331c-bb6a-475a-847f-40b25351917e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00a2685754870c267a4a3c2ce6f1b0ba23302bd2a2d7ea9030765d1d4f28c4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgn6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wn7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.356350 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9vsj7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d33f9d63-34d7-4722-95d5-13bdcb32b99a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7\\\\n2026-03-20T08:58:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_25c012a0-233e-40e1-80bb-fd663e59bcc7 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2zf2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9vsj7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.371432 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7322ac31-c6c4-43b9-8b3c-60ff84742545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bc8a5fb63feaa7fac2ab25e38eed542c11bf005da389d7f5b68874b5178dcd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cgm2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-snzb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.383632 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-z4b65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7748c732-655d-42c5-9dee-f7c69315471b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d51f4416a624ebeb3b8bb48c007ce4db4fa31660ea8f92f9deb4327acfe3693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tvgxr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-z4b65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.398575 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.413277 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://075b2db0b8c11f1eda4c6f5b7daecc53cd6d4d01900f65352231e0af4ac2b591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cced5baa687182f1330492445b8ccf64499801ffd7d4dce1a928dd63016d7756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:04 crc kubenswrapper[4823]: I0320 09:00:04.429607 4823 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5e3438faa1b0deef440f26a2df416ed22c04448cf74824952797f548ee4bc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:04Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:05 crc kubenswrapper[4823]: I0320 09:00:05.096484 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:05 crc kubenswrapper[4823]: I0320 09:00:05.096597 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:05 crc kubenswrapper[4823]: I0320 09:00:05.096885 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:05 crc kubenswrapper[4823]: I0320 09:00:05.096957 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:05 crc kubenswrapper[4823]: E0320 09:00:05.097170 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:05 crc kubenswrapper[4823]: E0320 09:00:05.097347 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:05 crc kubenswrapper[4823]: E0320 09:00:05.097523 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:05 crc kubenswrapper[4823]: E0320 09:00:05.097694 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.412997 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.413064 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.413082 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.413111 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.413130 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T09:00:06Z","lastTransitionTime":"2026-03-20T09:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 09:00:06 crc kubenswrapper[4823]: E0320 09:00:06.430097 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:06Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.434620 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.434712 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.434795 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.434863 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.434899 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T09:00:06Z","lastTransitionTime":"2026-03-20T09:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 09:00:06 crc kubenswrapper[4823]: E0320 09:00:06.452402 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:06Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.456706 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.456815 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.456842 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.456872 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.456896 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T09:00:06Z","lastTransitionTime":"2026-03-20T09:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 09:00:06 crc kubenswrapper[4823]: E0320 09:00:06.477702 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:06Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.482095 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.482136 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.482151 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.482173 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.482187 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T09:00:06Z","lastTransitionTime":"2026-03-20T09:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 09:00:06 crc kubenswrapper[4823]: E0320 09:00:06.499361 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:06Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.503597 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.503663 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.503683 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.503706 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 09:00:06 crc kubenswrapper[4823]: I0320 09:00:06.503723 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T09:00:06Z","lastTransitionTime":"2026-03-20T09:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 09:00:06 crc kubenswrapper[4823]: E0320 09:00:06.524150 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T09:00:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8f9fe74c-818f-44ae-ba11-a9422f2d82f7\\\",\\\"systemUUID\\\":\\\"62ada254-c183-4bbc-aaa5-3bc10b9c932c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T09:00:06Z is after 2025-08-24T17:21:41Z" Mar 20 09:00:06 crc kubenswrapper[4823]: E0320 09:00:06.524375 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 09:00:07 crc kubenswrapper[4823]: I0320 09:00:07.096624 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:07 crc kubenswrapper[4823]: I0320 09:00:07.096726 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:07 crc kubenswrapper[4823]: I0320 09:00:07.096870 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:07 crc kubenswrapper[4823]: E0320 09:00:07.098714 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:07 crc kubenswrapper[4823]: I0320 09:00:07.099153 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:07 crc kubenswrapper[4823]: E0320 09:00:07.099524 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:07 crc kubenswrapper[4823]: E0320 09:00:07.099853 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:07 crc kubenswrapper[4823]: E0320 09:00:07.100140 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:09 crc kubenswrapper[4823]: I0320 09:00:09.096668 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:09 crc kubenswrapper[4823]: I0320 09:00:09.096782 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:09 crc kubenswrapper[4823]: E0320 09:00:09.096815 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:09 crc kubenswrapper[4823]: I0320 09:00:09.096825 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:09 crc kubenswrapper[4823]: I0320 09:00:09.096884 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:09 crc kubenswrapper[4823]: E0320 09:00:09.097040 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:09 crc kubenswrapper[4823]: E0320 09:00:09.097436 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:09 crc kubenswrapper[4823]: E0320 09:00:09.097696 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:09 crc kubenswrapper[4823]: E0320 09:00:09.213706 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:11 crc kubenswrapper[4823]: I0320 09:00:11.096703 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:11 crc kubenswrapper[4823]: E0320 09:00:11.096946 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:11 crc kubenswrapper[4823]: I0320 09:00:11.097003 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:11 crc kubenswrapper[4823]: I0320 09:00:11.097121 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:11 crc kubenswrapper[4823]: E0320 09:00:11.097224 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:11 crc kubenswrapper[4823]: E0320 09:00:11.097321 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:11 crc kubenswrapper[4823]: I0320 09:00:11.097902 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:11 crc kubenswrapper[4823]: E0320 09:00:11.098059 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:13 crc kubenswrapper[4823]: I0320 09:00:13.097177 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:13 crc kubenswrapper[4823]: I0320 09:00:13.097331 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:13 crc kubenswrapper[4823]: E0320 09:00:13.097372 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:13 crc kubenswrapper[4823]: I0320 09:00:13.097405 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:13 crc kubenswrapper[4823]: I0320 09:00:13.097345 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:13 crc kubenswrapper[4823]: E0320 09:00:13.097602 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:13 crc kubenswrapper[4823]: E0320 09:00:13.097890 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:13 crc kubenswrapper[4823]: E0320 09:00:13.097971 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:14 crc kubenswrapper[4823]: I0320 09:00:14.178506 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-z4b65" podStartSLOduration=139.178485871 podStartE2EDuration="2m19.178485871s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:14.158617835 +0000 UTC m=+180.844514219" watchObservedRunningTime="2026-03-20 09:00:14.178485871 +0000 UTC m=+180.864382235" Mar 20 09:00:14 crc kubenswrapper[4823]: E0320 09:00:14.215506 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:14 crc kubenswrapper[4823]: I0320 09:00:14.244509 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7wn7x" podStartSLOduration=139.244467833 podStartE2EDuration="2m19.244467833s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:14.226945453 +0000 UTC m=+180.912841837" watchObservedRunningTime="2026-03-20 09:00:14.244467833 +0000 UTC m=+180.930364207" Mar 20 09:00:14 crc kubenswrapper[4823]: I0320 09:00:14.244763 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-9vsj7" podStartSLOduration=139.244733812 podStartE2EDuration="2m19.244733812s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:14.24469538 +0000 UTC m=+180.930591774" watchObservedRunningTime="2026-03-20 09:00:14.244733812 +0000 UTC m=+180.930630186" Mar 20 09:00:14 crc kubenswrapper[4823]: I0320 09:00:14.279564 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podStartSLOduration=139.279532656 podStartE2EDuration="2m19.279532656s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:14.259867297 +0000 UTC m=+180.945763671" watchObservedRunningTime="2026-03-20 09:00:14.279532656 +0000 UTC m=+180.965429050" Mar 20 09:00:14 crc kubenswrapper[4823]: I0320 09:00:14.280037 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=64.280032242 podStartE2EDuration="1m4.280032242s" podCreationTimestamp="2026-03-20 08:59:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:14.279196985 +0000 UTC m=+180.965093369" watchObservedRunningTime="2026-03-20 09:00:14.280032242 +0000 UTC m=+180.965928616" Mar 20 09:00:14 crc kubenswrapper[4823]: I0320 09:00:14.292128 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=88.292109528 podStartE2EDuration="1m28.292109528s" podCreationTimestamp="2026-03-20 08:58:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:14.291790518 +0000 UTC m=+180.977686902" watchObservedRunningTime="2026-03-20 09:00:14.292109528 +0000 UTC m=+180.978005902" Mar 20 09:00:14 crc kubenswrapper[4823]: I0320 09:00:14.391279 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=85.391254333 podStartE2EDuration="1m25.391254333s" podCreationTimestamp="2026-03-20 08:58:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:14.390834909 +0000 UTC m=+181.076731293" watchObservedRunningTime="2026-03-20 09:00:14.391254333 +0000 UTC m=+181.077150707" Mar 20 09:00:14 crc kubenswrapper[4823]: I0320 09:00:14.407284 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cx7cq" podStartSLOduration=138.407251635 podStartE2EDuration="2m18.407251635s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:14.406396908 +0000 UTC m=+181.092293282" watchObservedRunningTime="2026-03-20 09:00:14.407251635 +0000 UTC m=+181.093148009" Mar 20 09:00:14 crc kubenswrapper[4823]: I0320 09:00:14.422658 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=53.422630328 podStartE2EDuration="53.422630328s" podCreationTimestamp="2026-03-20 08:59:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:14.42086019 +0000 UTC m=+181.106756554" watchObservedRunningTime="2026-03-20 09:00:14.422630328 +0000 UTC m=+181.108526702" Mar 20 09:00:14 crc kubenswrapper[4823]: I0320 09:00:14.472794 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=84.472772893 podStartE2EDuration="1m24.472772893s" podCreationTimestamp="2026-03-20 08:58:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:14.457309197 +0000 UTC m=+181.143205591" watchObservedRunningTime="2026-03-20 09:00:14.472772893 +0000 UTC m=+181.158669267" Mar 20 09:00:15 crc kubenswrapper[4823]: I0320 09:00:15.096954 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:15 crc kubenswrapper[4823]: I0320 09:00:15.097086 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:15 crc kubenswrapper[4823]: I0320 09:00:15.097107 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:15 crc kubenswrapper[4823]: E0320 09:00:15.097246 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:15 crc kubenswrapper[4823]: E0320 09:00:15.097395 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:15 crc kubenswrapper[4823]: E0320 09:00:15.097497 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:15 crc kubenswrapper[4823]: I0320 09:00:15.097931 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:15 crc kubenswrapper[4823]: E0320 09:00:15.098058 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.098572 4823 scope.go:117] "RemoveContainer" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 09:00:16 crc kubenswrapper[4823]: E0320 09:00:16.099055 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.828061 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.828133 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.828149 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.828182 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.828196 4823 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T09:00:16Z","lastTransitionTime":"2026-03-20T09:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.928390 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-fb4mg" podStartSLOduration=141.92837195 podStartE2EDuration="2m21.92837195s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:14.491306056 +0000 UTC m=+181.177202430" watchObservedRunningTime="2026-03-20 09:00:16.92837195 +0000 UTC m=+183.614268334" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.928649 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr"] Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.929154 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.931621 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.931724 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.931979 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.935112 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.974415 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c4b7a110-905e-46a2-8459-12878b767fad-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.974521 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4b7a110-905e-46a2-8459-12878b767fad-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.974723 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b7a110-905e-46a2-8459-12878b767fad-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.974919 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c4b7a110-905e-46a2-8459-12878b767fad-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:16 crc kubenswrapper[4823]: I0320 09:00:16.974995 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c4b7a110-905e-46a2-8459-12878b767fad-service-ca\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.075980 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c4b7a110-905e-46a2-8459-12878b767fad-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.076024 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c4b7a110-905e-46a2-8459-12878b767fad-service-ca\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.076049 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c4b7a110-905e-46a2-8459-12878b767fad-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.076072 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4b7a110-905e-46a2-8459-12878b767fad-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.076098 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b7a110-905e-46a2-8459-12878b767fad-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.077002 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c4b7a110-905e-46a2-8459-12878b767fad-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.077187 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c4b7a110-905e-46a2-8459-12878b767fad-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.077413 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c4b7a110-905e-46a2-8459-12878b767fad-service-ca\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.082411 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b7a110-905e-46a2-8459-12878b767fad-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.096298 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.096320 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.096343 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.096316 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:17 crc kubenswrapper[4823]: E0320 09:00:17.096429 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:17 crc kubenswrapper[4823]: E0320 09:00:17.096687 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:17 crc kubenswrapper[4823]: E0320 09:00:17.096680 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:17 crc kubenswrapper[4823]: E0320 09:00:17.096773 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.116298 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4b7a110-905e-46a2-8459-12878b767fad-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-q2mrr\" (UID: \"c4b7a110-905e-46a2-8459-12878b767fad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.143733 4823 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.153809 4823 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 20 09:00:17 crc kubenswrapper[4823]: I0320 09:00:17.250289 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" Mar 20 09:00:18 crc kubenswrapper[4823]: I0320 09:00:18.036874 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" event={"ID":"c4b7a110-905e-46a2-8459-12878b767fad","Type":"ContainerStarted","Data":"ccbb9d8b6df8042684813db7d98098c775d29723fdf87056420aa934964205fb"} Mar 20 09:00:18 crc kubenswrapper[4823]: I0320 09:00:18.036930 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" event={"ID":"c4b7a110-905e-46a2-8459-12878b767fad","Type":"ContainerStarted","Data":"23fdd3c8ca378c92c3038133185904f6f750b0c11ec805c28ce7f63a8d49a283"} Mar 20 09:00:18 crc kubenswrapper[4823]: I0320 09:00:18.059775 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q2mrr" podStartSLOduration=143.0597144 podStartE2EDuration="2m23.0597144s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:18.059656618 +0000 UTC m=+184.745553072" watchObservedRunningTime="2026-03-20 09:00:18.0597144 +0000 UTC m=+184.745610814" Mar 20 09:00:19 crc kubenswrapper[4823]: I0320 09:00:19.096587 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:19 crc kubenswrapper[4823]: I0320 09:00:19.096667 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:19 crc kubenswrapper[4823]: I0320 09:00:19.096732 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:19 crc kubenswrapper[4823]: I0320 09:00:19.096866 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:19 crc kubenswrapper[4823]: E0320 09:00:19.096860 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:19 crc kubenswrapper[4823]: E0320 09:00:19.097019 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:19 crc kubenswrapper[4823]: E0320 09:00:19.097213 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:19 crc kubenswrapper[4823]: E0320 09:00:19.097315 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:19 crc kubenswrapper[4823]: E0320 09:00:19.217608 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:21 crc kubenswrapper[4823]: I0320 09:00:21.096419 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:21 crc kubenswrapper[4823]: I0320 09:00:21.096469 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:21 crc kubenswrapper[4823]: I0320 09:00:21.096491 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:21 crc kubenswrapper[4823]: I0320 09:00:21.096444 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:21 crc kubenswrapper[4823]: E0320 09:00:21.096715 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:21 crc kubenswrapper[4823]: E0320 09:00:21.096875 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:21 crc kubenswrapper[4823]: E0320 09:00:21.096989 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:21 crc kubenswrapper[4823]: E0320 09:00:21.097055 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:23 crc kubenswrapper[4823]: I0320 09:00:23.096964 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:23 crc kubenswrapper[4823]: I0320 09:00:23.097040 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:23 crc kubenswrapper[4823]: I0320 09:00:23.097058 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:23 crc kubenswrapper[4823]: I0320 09:00:23.096992 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:23 crc kubenswrapper[4823]: E0320 09:00:23.097199 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:23 crc kubenswrapper[4823]: E0320 09:00:23.097332 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:23 crc kubenswrapper[4823]: E0320 09:00:23.097520 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:23 crc kubenswrapper[4823]: E0320 09:00:23.097611 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:24 crc kubenswrapper[4823]: E0320 09:00:24.218920 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:25 crc kubenswrapper[4823]: I0320 09:00:25.096166 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:25 crc kubenswrapper[4823]: I0320 09:00:25.096234 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:25 crc kubenswrapper[4823]: E0320 09:00:25.096298 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:25 crc kubenswrapper[4823]: E0320 09:00:25.096450 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:25 crc kubenswrapper[4823]: I0320 09:00:25.096494 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:25 crc kubenswrapper[4823]: E0320 09:00:25.096540 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:25 crc kubenswrapper[4823]: I0320 09:00:25.096580 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:25 crc kubenswrapper[4823]: E0320 09:00:25.096647 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:27 crc kubenswrapper[4823]: I0320 09:00:27.069520 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9vsj7_d33f9d63-34d7-4722-95d5-13bdcb32b99a/kube-multus/1.log" Mar 20 09:00:27 crc kubenswrapper[4823]: I0320 09:00:27.070966 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9vsj7_d33f9d63-34d7-4722-95d5-13bdcb32b99a/kube-multus/0.log" Mar 20 09:00:27 crc kubenswrapper[4823]: I0320 09:00:27.071036 4823 generic.go:334] "Generic (PLEG): container finished" podID="d33f9d63-34d7-4722-95d5-13bdcb32b99a" containerID="78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b" exitCode=1 Mar 20 09:00:27 crc kubenswrapper[4823]: I0320 09:00:27.071079 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9vsj7" event={"ID":"d33f9d63-34d7-4722-95d5-13bdcb32b99a","Type":"ContainerDied","Data":"78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b"} Mar 20 09:00:27 crc kubenswrapper[4823]: I0320 09:00:27.071127 4823 scope.go:117] "RemoveContainer" containerID="05ed79710c9fcc9cd20bacc826370e1a0598998e211e357d32df32f2c985eba8" Mar 20 09:00:27 crc kubenswrapper[4823]: I0320 09:00:27.071862 4823 scope.go:117] "RemoveContainer" containerID="78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b" Mar 20 09:00:27 crc kubenswrapper[4823]: E0320 09:00:27.072271 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-9vsj7_openshift-multus(d33f9d63-34d7-4722-95d5-13bdcb32b99a)\"" pod="openshift-multus/multus-9vsj7" podUID="d33f9d63-34d7-4722-95d5-13bdcb32b99a" Mar 20 09:00:27 crc kubenswrapper[4823]: I0320 09:00:27.101554 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:27 crc kubenswrapper[4823]: I0320 09:00:27.101598 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:27 crc kubenswrapper[4823]: I0320 09:00:27.101564 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:27 crc kubenswrapper[4823]: E0320 09:00:27.101693 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:27 crc kubenswrapper[4823]: I0320 09:00:27.101733 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:27 crc kubenswrapper[4823]: E0320 09:00:27.101939 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:27 crc kubenswrapper[4823]: E0320 09:00:27.102370 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:27 crc kubenswrapper[4823]: E0320 09:00:27.102497 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:27 crc kubenswrapper[4823]: I0320 09:00:27.102765 4823 scope.go:117] "RemoveContainer" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 09:00:27 crc kubenswrapper[4823]: E0320 09:00:27.103033 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2x54k_openshift-ovn-kubernetes(1bed1862-5bcd-46fe-bc84-2dbab416086b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" Mar 20 09:00:28 crc kubenswrapper[4823]: I0320 09:00:28.081076 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9vsj7_d33f9d63-34d7-4722-95d5-13bdcb32b99a/kube-multus/1.log" Mar 20 09:00:29 crc kubenswrapper[4823]: I0320 09:00:29.096929 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:29 crc kubenswrapper[4823]: I0320 09:00:29.097037 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:29 crc kubenswrapper[4823]: E0320 09:00:29.097075 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:29 crc kubenswrapper[4823]: I0320 09:00:29.097104 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:29 crc kubenswrapper[4823]: I0320 09:00:29.097027 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:29 crc kubenswrapper[4823]: E0320 09:00:29.097208 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:29 crc kubenswrapper[4823]: E0320 09:00:29.097352 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:29 crc kubenswrapper[4823]: E0320 09:00:29.097417 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:29 crc kubenswrapper[4823]: E0320 09:00:29.231205 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:31 crc kubenswrapper[4823]: I0320 09:00:31.096206 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:31 crc kubenswrapper[4823]: I0320 09:00:31.096222 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:31 crc kubenswrapper[4823]: I0320 09:00:31.096311 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:31 crc kubenswrapper[4823]: E0320 09:00:31.096522 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:31 crc kubenswrapper[4823]: I0320 09:00:31.096570 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:31 crc kubenswrapper[4823]: E0320 09:00:31.096773 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:31 crc kubenswrapper[4823]: E0320 09:00:31.096912 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:31 crc kubenswrapper[4823]: E0320 09:00:31.097109 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:33 crc kubenswrapper[4823]: I0320 09:00:33.097102 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:33 crc kubenswrapper[4823]: I0320 09:00:33.097189 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:33 crc kubenswrapper[4823]: I0320 09:00:33.097152 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:33 crc kubenswrapper[4823]: I0320 09:00:33.097111 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:33 crc kubenswrapper[4823]: E0320 09:00:33.097681 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:33 crc kubenswrapper[4823]: E0320 09:00:33.097875 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:33 crc kubenswrapper[4823]: E0320 09:00:33.098066 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:33 crc kubenswrapper[4823]: E0320 09:00:33.098245 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:34 crc kubenswrapper[4823]: E0320 09:00:34.232711 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:35 crc kubenswrapper[4823]: I0320 09:00:35.096264 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:35 crc kubenswrapper[4823]: I0320 09:00:35.096323 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:35 crc kubenswrapper[4823]: I0320 09:00:35.096340 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:35 crc kubenswrapper[4823]: I0320 09:00:35.096260 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:35 crc kubenswrapper[4823]: E0320 09:00:35.096535 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:35 crc kubenswrapper[4823]: E0320 09:00:35.096734 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:35 crc kubenswrapper[4823]: E0320 09:00:35.096913 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:35 crc kubenswrapper[4823]: E0320 09:00:35.097060 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:37 crc kubenswrapper[4823]: I0320 09:00:37.096980 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:37 crc kubenswrapper[4823]: I0320 09:00:37.097032 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:37 crc kubenswrapper[4823]: I0320 09:00:37.096988 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:37 crc kubenswrapper[4823]: E0320 09:00:37.097098 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:37 crc kubenswrapper[4823]: E0320 09:00:37.097215 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:37 crc kubenswrapper[4823]: I0320 09:00:37.097276 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:37 crc kubenswrapper[4823]: E0320 09:00:37.097379 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:37 crc kubenswrapper[4823]: E0320 09:00:37.097470 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:39 crc kubenswrapper[4823]: I0320 09:00:39.097199 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:39 crc kubenswrapper[4823]: I0320 09:00:39.097305 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:39 crc kubenswrapper[4823]: E0320 09:00:39.097414 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:39 crc kubenswrapper[4823]: I0320 09:00:39.097491 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:39 crc kubenswrapper[4823]: I0320 09:00:39.097214 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:39 crc kubenswrapper[4823]: E0320 09:00:39.097628 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:39 crc kubenswrapper[4823]: E0320 09:00:39.097774 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:39 crc kubenswrapper[4823]: E0320 09:00:39.097880 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:39 crc kubenswrapper[4823]: E0320 09:00:39.278353 4823 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:41 crc kubenswrapper[4823]: I0320 09:00:41.096665 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:41 crc kubenswrapper[4823]: I0320 09:00:41.096718 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:41 crc kubenswrapper[4823]: I0320 09:00:41.096950 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:41 crc kubenswrapper[4823]: I0320 09:00:41.096996 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:41 crc kubenswrapper[4823]: E0320 09:00:41.097106 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:41 crc kubenswrapper[4823]: I0320 09:00:41.097214 4823 scope.go:117] "RemoveContainer" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 09:00:41 crc kubenswrapper[4823]: E0320 09:00:41.097422 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:41 crc kubenswrapper[4823]: E0320 09:00:41.097518 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:41 crc kubenswrapper[4823]: E0320 09:00:41.097600 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:41 crc kubenswrapper[4823]: I0320 09:00:41.098845 4823 scope.go:117] "RemoveContainer" containerID="78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b" Mar 20 09:00:42 crc kubenswrapper[4823]: I0320 09:00:42.030015 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jzzk2"] Mar 20 09:00:42 crc kubenswrapper[4823]: I0320 09:00:42.030487 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:42 crc kubenswrapper[4823]: E0320 09:00:42.030591 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:42 crc kubenswrapper[4823]: I0320 09:00:42.136790 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9vsj7_d33f9d63-34d7-4722-95d5-13bdcb32b99a/kube-multus/1.log" Mar 20 09:00:42 crc kubenswrapper[4823]: I0320 09:00:42.136901 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9vsj7" event={"ID":"d33f9d63-34d7-4722-95d5-13bdcb32b99a","Type":"ContainerStarted","Data":"2dae75ee51902fd1a5cb22fba350a24fe1325a6e77293756078944b96466dffe"} Mar 20 09:00:42 crc kubenswrapper[4823]: I0320 09:00:42.140830 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/3.log" Mar 20 09:00:42 crc kubenswrapper[4823]: I0320 09:00:42.143526 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerStarted","Data":"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1"} Mar 20 09:00:42 crc kubenswrapper[4823]: I0320 09:00:42.144044 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 09:00:42 crc kubenswrapper[4823]: I0320 09:00:42.189127 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podStartSLOduration=167.189105784 podStartE2EDuration="2m47.189105784s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:42.188419235 +0000 UTC m=+208.874315659" watchObservedRunningTime="2026-03-20 09:00:42.189105784 +0000 UTC m=+208.875002178" Mar 20 09:00:43 crc kubenswrapper[4823]: I0320 09:00:43.097181 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:43 crc kubenswrapper[4823]: I0320 09:00:43.097277 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:43 crc kubenswrapper[4823]: I0320 09:00:43.097322 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:43 crc kubenswrapper[4823]: E0320 09:00:43.097496 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:43 crc kubenswrapper[4823]: I0320 09:00:43.097522 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:43 crc kubenswrapper[4823]: E0320 09:00:43.097691 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jzzk2" podUID="316f49ed-2e3c-42b7-a133-a36c3074810e" Mar 20 09:00:43 crc kubenswrapper[4823]: E0320 09:00:43.097853 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:43 crc kubenswrapper[4823]: E0320 09:00:43.097935 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:45 crc kubenswrapper[4823]: I0320 09:00:45.096172 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:45 crc kubenswrapper[4823]: I0320 09:00:45.096181 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:45 crc kubenswrapper[4823]: I0320 09:00:45.096221 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:45 crc kubenswrapper[4823]: I0320 09:00:45.096230 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:45 crc kubenswrapper[4823]: I0320 09:00:45.099257 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 20 09:00:45 crc kubenswrapper[4823]: I0320 09:00:45.101168 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 20 09:00:45 crc kubenswrapper[4823]: I0320 09:00:45.101425 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 20 09:00:45 crc kubenswrapper[4823]: I0320 09:00:45.101655 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 20 09:00:45 crc kubenswrapper[4823]: I0320 09:00:45.101777 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 20 09:00:45 crc kubenswrapper[4823]: I0320 09:00:45.103807 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.072661 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.074600 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.173944 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.174072 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:47 crc kubenswrapper[4823]: E0320 09:00:47.174153 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:02:49.174110747 +0000 UTC m=+335.860007151 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.174215 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.174298 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.181693 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.181719 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.189551 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.224570 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.242637 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.270102 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.275676 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.280422 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/316f49ed-2e3c-42b7-a133-a36c3074810e-metrics-certs\") pod \"network-metrics-daemon-jzzk2\" (UID: \"316f49ed-2e3c-42b7-a133-a36c3074810e\") " pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.455961 4823 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.493460 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.493875 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.496245 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.501402 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.501664 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.501843 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.501881 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-lz4n9"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.502033 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.502146 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.502324 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sswrk"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.502524 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.502729 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-49lkt"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.503158 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.503315 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.503365 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.503576 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.503702 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.503894 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.504193 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.509164 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.509270 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.509497 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.509522 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.509799 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.509957 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.510094 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.510231 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.510395 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.510525 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.510656 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.510813 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.511000 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.517165 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.517476 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.517760 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.518028 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.518185 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.518340 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.518558 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.518583 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-grjk6"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.518886 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.518920 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.519186 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.519312 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.519547 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.521452 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.521776 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.521936 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.521969 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-grjk6" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.522330 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.522520 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.523888 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.524062 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.524221 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.528451 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.553629 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.554094 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.568786 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jzzk2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.569147 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.569667 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.570409 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qm6tg"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.570883 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mkqpg"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.571356 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.571580 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.571631 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.571824 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.571917 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.573151 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.573703 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.574703 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.575793 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.578927 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.578970 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8777e543-c661-4a72-9221-e4d398222993-etcd-service-ca\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579015 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-etcd-client\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579045 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w8b2\" (UniqueName: \"kubernetes.io/projected/65497b00-f979-4bad-a290-0af7be716ace-kube-api-access-7w8b2\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579068 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ce7e9b8-5c85-4ea2-ad33-14bdf64be778-metrics-tls\") pod \"dns-operator-744455d44c-mkqpg\" (UID: \"5ce7e9b8-5c85-4ea2-ad33-14bdf64be778\") " pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579090 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65497b00-f979-4bad-a290-0af7be716ace-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579112 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579133 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4256007-3ebf-4991-89b5-83eb7104b8a6-metrics-tls\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579152 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/65497b00-f979-4bad-a290-0af7be716ace-audit-dir\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579174 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-encryption-config\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579196 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-audit-dir\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579220 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-oauth-serving-cert\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579242 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zqlv\" (UniqueName: \"kubernetes.io/projected/dbf97305-8126-4377-801a-3b57bea03ea3-kube-api-access-5zqlv\") pod \"cluster-samples-operator-665b6dd947-7b46d\" (UID: \"dbf97305-8126-4377-801a-3b57bea03ea3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579264 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrrcv\" (UniqueName: \"kubernetes.io/projected/98f13ee0-723b-4ef5-851b-2665009a9d47-kube-api-access-vrrcv\") pod \"downloads-7954f5f757-grjk6\" (UID: \"98f13ee0-723b-4ef5-851b-2665009a9d47\") " pod="openshift-console/downloads-7954f5f757-grjk6" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579284 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-etcd-serving-ca\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579304 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/65497b00-f979-4bad-a290-0af7be716ace-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579330 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-oauth-config\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579351 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579375 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-console-config\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579397 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8777e543-c661-4a72-9221-e4d398222993-etcd-ca\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579416 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/65497b00-f979-4bad-a290-0af7be716ace-audit-policies\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579436 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/65497b00-f979-4bad-a290-0af7be716ace-encryption-config\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579462 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-config\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579484 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c4256007-3ebf-4991-89b5-83eb7104b8a6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579506 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c4256007-3ebf-4991-89b5-83eb7104b8a6-trusted-ca\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579526 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-client-ca\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579547 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vbw6\" (UniqueName: \"kubernetes.io/projected/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-kube-api-access-8vbw6\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579568 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-serving-cert\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579589 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r22wq\" (UniqueName: \"kubernetes.io/projected/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-kube-api-access-r22wq\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579611 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtqbq\" (UniqueName: \"kubernetes.io/projected/8777e543-c661-4a72-9221-e4d398222993-kube-api-access-gtqbq\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579636 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-node-pullsecrets\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579659 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65497b00-f979-4bad-a290-0af7be716ace-serving-cert\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579684 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-trusted-ca-bundle\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579706 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-serving-cert\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579728 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-service-ca\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579766 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/65497b00-f979-4bad-a290-0af7be716ace-etcd-client\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579785 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-serving-cert\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579809 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gczzv\" (UniqueName: \"kubernetes.io/projected/c4256007-3ebf-4991-89b5-83eb7104b8a6-kube-api-access-gczzv\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579831 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-config\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579852 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46s55\" (UniqueName: \"kubernetes.io/projected/e104c875-7420-45b3-b675-a63b92c454e4-kube-api-access-46s55\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579880 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-images\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579905 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8777e543-c661-4a72-9221-e4d398222993-serving-cert\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579925 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8777e543-c661-4a72-9221-e4d398222993-config\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.579947 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dbf97305-8126-4377-801a-3b57bea03ea3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7b46d\" (UID: \"dbf97305-8126-4377-801a-3b57bea03ea3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.580091 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8777e543-c661-4a72-9221-e4d398222993-etcd-client\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.580117 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-config\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.580142 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g6cr\" (UniqueName: \"kubernetes.io/projected/5ce7e9b8-5c85-4ea2-ad33-14bdf64be778-kube-api-access-7g6cr\") pod \"dns-operator-744455d44c-mkqpg\" (UID: \"5ce7e9b8-5c85-4ea2-ad33-14bdf64be778\") " pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.580195 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k94q\" (UniqueName: \"kubernetes.io/projected/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-kube-api-access-2k94q\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.580227 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-audit\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.580255 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-image-import-ca\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.580939 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.581366 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.581458 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.581524 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.581658 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.581857 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.582020 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.582127 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.582203 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.582275 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.582422 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.582498 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.582572 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.582640 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.582780 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.582862 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.583281 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.583455 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.583559 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.583637 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-lndmn"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.583998 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.584349 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.584577 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.589396 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.589513 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.589711 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.589801 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.589830 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.594840 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.594990 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.595108 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.595326 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.595551 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.600763 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.606210 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.609060 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.609383 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.609499 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.609614 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.609842 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.612046 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwrsf"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.612605 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.613279 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.613680 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.616971 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zdsvr"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.617456 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.618308 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.618467 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.618845 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.620449 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.620569 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.620558 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.620895 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.621011 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.621113 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.621133 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.621284 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.621333 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.621453 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.626327 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.626678 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.626922 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.627071 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.628446 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qsxgq"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.628846 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.628871 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.640796 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.641369 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.641980 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.642065 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.644011 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.644221 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.644254 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.644582 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.651250 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.652018 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.653663 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.655386 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.655617 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.656704 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.663659 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.664348 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.665139 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.665973 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.670354 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wd9j2"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.671107 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.676638 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.676862 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.678791 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rfbmh"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.679573 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.680135 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7qskq"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.680553 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.680701 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.680871 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566620-4lfkt"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.680983 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.681580 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566620-4lfkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.680936 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.684972 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-serving-cert\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685008 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/249a7568-39b3-4332-a847-67a1b08d00bc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tbm9g\" (UID: \"249a7568-39b3-4332-a847-67a1b08d00bc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685033 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r22wq\" (UniqueName: \"kubernetes.io/projected/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-kube-api-access-r22wq\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685052 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtqbq\" (UniqueName: \"kubernetes.io/projected/8777e543-c661-4a72-9221-e4d398222993-kube-api-access-gtqbq\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685068 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-node-pullsecrets\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685085 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8580a746-c68f-4b57-8f69-f31fb855b785-config\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685107 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-trusted-ca-bundle\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685122 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65497b00-f979-4bad-a290-0af7be716ace-serving-cert\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685141 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685158 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljfjw\" (UniqueName: \"kubernetes.io/projected/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-kube-api-access-ljfjw\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685178 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xw7gs\" (UID: \"2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685218 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-serving-cert\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685239 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-service-ca\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685254 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/65497b00-f979-4bad-a290-0af7be716ace-etcd-client\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685269 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-serving-cert\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685284 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8580a746-c68f-4b57-8f69-f31fb855b785-auth-proxy-config\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685305 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b3679cb-16fd-4774-88a5-15d92fe16a43-proxy-tls\") pod \"machine-config-controller-84d6567774-sx7sz\" (UID: \"5b3679cb-16fd-4774-88a5-15d92fe16a43\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685323 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvmdp\" (UniqueName: \"kubernetes.io/projected/4da55054-ed9d-4d06-a94c-5f225541f7ea-kube-api-access-pvmdp\") pod \"control-plane-machine-set-operator-78cbb6b69f-qst64\" (UID: \"4da55054-ed9d-4d06-a94c-5f225541f7ea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685343 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gczzv\" (UniqueName: \"kubernetes.io/projected/c4256007-3ebf-4991-89b5-83eb7104b8a6-kube-api-access-gczzv\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685361 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-config\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685380 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685397 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/249a7568-39b3-4332-a847-67a1b08d00bc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tbm9g\" (UID: \"249a7568-39b3-4332-a847-67a1b08d00bc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685415 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-images\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685439 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46s55\" (UniqueName: \"kubernetes.io/projected/e104c875-7420-45b3-b675-a63b92c454e4-kube-api-access-46s55\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685460 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b3679cb-16fd-4774-88a5-15d92fe16a43-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sx7sz\" (UID: \"5b3679cb-16fd-4774-88a5-15d92fe16a43\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685480 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr5hz\" (UniqueName: \"kubernetes.io/projected/77917c9c-14bd-4653-ac86-571eda870a4c-kube-api-access-dr5hz\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685500 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8777e543-c661-4a72-9221-e4d398222993-serving-cert\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685517 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8777e543-c661-4a72-9221-e4d398222993-config\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685533 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/71d31547-2185-4113-a21e-2f0f2ae4423c-stats-auth\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685553 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dbf97305-8126-4377-801a-3b57bea03ea3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7b46d\" (UID: \"dbf97305-8126-4377-801a-3b57bea03ea3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685575 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8580a746-c68f-4b57-8f69-f31fb855b785-machine-approver-tls\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685592 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2-config\") pod \"kube-apiserver-operator-766d6c64bb-xw7gs\" (UID: \"2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685608 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xw7gs\" (UID: \"2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685626 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqbg2\" (UniqueName: \"kubernetes.io/projected/45124dc2-68b5-411c-9826-be34d809a370-kube-api-access-zqbg2\") pod \"openshift-apiserver-operator-796bbdcf4f-jbsnd\" (UID: \"45124dc2-68b5-411c-9826-be34d809a370\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685630 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685654 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8777e543-c661-4a72-9221-e4d398222993-etcd-client\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685677 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71d31547-2185-4113-a21e-2f0f2ae4423c-service-ca-bundle\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685694 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-config\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685711 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g6cr\" (UniqueName: \"kubernetes.io/projected/5ce7e9b8-5c85-4ea2-ad33-14bdf64be778-kube-api-access-7g6cr\") pod \"dns-operator-744455d44c-mkqpg\" (UID: \"5ce7e9b8-5c85-4ea2-ad33-14bdf64be778\") " pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685757 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/71d31547-2185-4113-a21e-2f0f2ae4423c-metrics-certs\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685774 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77917c9c-14bd-4653-ac86-571eda870a4c-serving-cert\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685791 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/77917c9c-14bd-4653-ac86-571eda870a4c-trusted-ca\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685811 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k94q\" (UniqueName: \"kubernetes.io/projected/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-kube-api-access-2k94q\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685825 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-audit\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685841 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-image-import-ca\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685858 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bdc9b52-2233-463b-ab15-444987a74673-serving-cert\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685875 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4da55054-ed9d-4d06-a94c-5f225541f7ea-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qst64\" (UID: \"4da55054-ed9d-4d06-a94c-5f225541f7ea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685889 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-etcd-client\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685908 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8777e543-c661-4a72-9221-e4d398222993-etcd-service-ca\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685926 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w8b2\" (UniqueName: \"kubernetes.io/projected/65497b00-f979-4bad-a290-0af7be716ace-kube-api-access-7w8b2\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685943 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwhpt\" (UniqueName: \"kubernetes.io/projected/7bdc9b52-2233-463b-ab15-444987a74673-kube-api-access-gwhpt\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685966 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ce7e9b8-5c85-4ea2-ad33-14bdf64be778-metrics-tls\") pod \"dns-operator-744455d44c-mkqpg\" (UID: \"5ce7e9b8-5c85-4ea2-ad33-14bdf64be778\") " pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685982 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bdc9b52-2233-463b-ab15-444987a74673-service-ca-bundle\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.685999 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65497b00-f979-4bad-a290-0af7be716ace-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686014 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r7jj\" (UniqueName: \"kubernetes.io/projected/8580a746-c68f-4b57-8f69-f31fb855b785-kube-api-access-4r7jj\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686032 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686048 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4256007-3ebf-4991-89b5-83eb7104b8a6-metrics-tls\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686065 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/65497b00-f979-4bad-a290-0af7be716ace-audit-dir\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686083 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-encryption-config\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686097 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-audit-dir\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686111 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-etcd-serving-ca\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686126 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45124dc2-68b5-411c-9826-be34d809a370-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-jbsnd\" (UID: \"45124dc2-68b5-411c-9826-be34d809a370\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686144 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-oauth-serving-cert\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686159 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zqlv\" (UniqueName: \"kubernetes.io/projected/dbf97305-8126-4377-801a-3b57bea03ea3-kube-api-access-5zqlv\") pod \"cluster-samples-operator-665b6dd947-7b46d\" (UID: \"dbf97305-8126-4377-801a-3b57bea03ea3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686174 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrrcv\" (UniqueName: \"kubernetes.io/projected/98f13ee0-723b-4ef5-851b-2665009a9d47-kube-api-access-vrrcv\") pod \"downloads-7954f5f757-grjk6\" (UID: \"98f13ee0-723b-4ef5-851b-2665009a9d47\") " pod="openshift-console/downloads-7954f5f757-grjk6" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686190 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/65497b00-f979-4bad-a290-0af7be716ace-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686209 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh7xz\" (UniqueName: \"kubernetes.io/projected/f49fb4d8-ad57-4153-bd98-6b21c039b971-kube-api-access-dh7xz\") pod \"openshift-config-operator-7777fb866f-2jvr2\" (UID: \"f49fb4d8-ad57-4153-bd98-6b21c039b971\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686310 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686724 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686952 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.687008 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-node-pullsecrets\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.686224 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77917c9c-14bd-4653-ac86-571eda870a4c-config\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.687296 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jngfg\" (UniqueName: \"kubernetes.io/projected/5b3679cb-16fd-4774-88a5-15d92fe16a43-kube-api-access-jngfg\") pod \"machine-config-controller-84d6567774-sx7sz\" (UID: \"5b3679cb-16fd-4774-88a5-15d92fe16a43\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.687319 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-oauth-config\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.687345 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.687378 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bdc9b52-2233-463b-ab15-444987a74673-config\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.687403 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-console-config\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.687426 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8777e543-c661-4a72-9221-e4d398222993-etcd-ca\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.688705 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-config\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.688705 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-trusted-ca-bundle\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.689175 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hlgq9"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.690579 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-images\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.689899 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-config\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.690637 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-image-import-ca\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.690804 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/65497b00-f979-4bad-a290-0af7be716ace-audit-policies\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.690845 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/65497b00-f979-4bad-a290-0af7be716ace-encryption-config\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.690867 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-config\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.690889 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/249a7568-39b3-4332-a847-67a1b08d00bc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tbm9g\" (UID: \"249a7568-39b3-4332-a847-67a1b08d00bc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.690915 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45124dc2-68b5-411c-9826-be34d809a370-config\") pod \"openshift-apiserver-operator-796bbdcf4f-jbsnd\" (UID: \"45124dc2-68b5-411c-9826-be34d809a370\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.690933 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/71d31547-2185-4113-a21e-2f0f2ae4423c-default-certificate\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.690951 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.690973 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c4256007-3ebf-4991-89b5-83eb7104b8a6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.690991 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f49fb4d8-ad57-4153-bd98-6b21c039b971-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2jvr2\" (UID: \"f49fb4d8-ad57-4153-bd98-6b21c039b971\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.691008 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bdc9b52-2233-463b-ab15-444987a74673-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.691029 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vbw6\" (UniqueName: \"kubernetes.io/projected/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-kube-api-access-8vbw6\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.691047 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhbvm\" (UniqueName: \"kubernetes.io/projected/71d31547-2185-4113-a21e-2f0f2ae4423c-kube-api-access-qhbvm\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.691067 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c4256007-3ebf-4991-89b5-83eb7104b8a6-trusted-ca\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.691084 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-client-ca\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.691101 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f49fb4d8-ad57-4153-bd98-6b21c039b971-serving-cert\") pod \"openshift-config-operator-7777fb866f-2jvr2\" (UID: \"f49fb4d8-ad57-4153-bd98-6b21c039b971\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.692709 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.689812 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-audit\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.694044 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-audit-dir\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.694525 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8777e543-c661-4a72-9221-e4d398222993-config\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.694887 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8777e543-c661-4a72-9221-e4d398222993-etcd-service-ca\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.695027 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/65497b00-f979-4bad-a290-0af7be716ace-audit-dir\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.695117 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/65497b00-f979-4bad-a290-0af7be716ace-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.695231 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65497b00-f979-4bad-a290-0af7be716ace-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.695900 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-oauth-serving-cert\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.696411 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zdc26"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.696486 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-etcd-client\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.697118 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.697419 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8777e543-c661-4a72-9221-e4d398222993-serving-cert\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.697636 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4256007-3ebf-4991-89b5-83eb7104b8a6-metrics-tls\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.698855 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-etcd-serving-ca\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.699310 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ce7e9b8-5c85-4ea2-ad33-14bdf64be778-metrics-tls\") pod \"dns-operator-744455d44c-mkqpg\" (UID: \"5ce7e9b8-5c85-4ea2-ad33-14bdf64be778\") " pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.699723 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.701031 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-service-ca\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.701174 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8777e543-c661-4a72-9221-e4d398222993-etcd-ca\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.702766 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-console-config\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.702861 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c4256007-3ebf-4991-89b5-83eb7104b8a6-trusted-ca\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.703554 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.703706 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/65497b00-f979-4bad-a290-0af7be716ace-audit-policies\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.703962 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-client-ca\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.705953 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.708950 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-serving-cert\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.709724 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-oauth-config\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.712623 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p2lfv"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.712832 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.712898 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/65497b00-f979-4bad-a290-0af7be716ace-encryption-config\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.713839 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-49lkt"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.713861 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.714303 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.714317 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-lz4n9"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.714362 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.714861 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/65497b00-f979-4bad-a290-0af7be716ace-etcd-client\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.715029 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-encryption-config\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.715060 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-grjk6"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.715799 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dbf97305-8126-4377-801a-3b57bea03ea3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7b46d\" (UID: \"dbf97305-8126-4377-801a-3b57bea03ea3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.717167 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-config\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.717603 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-76jgq"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.718666 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.723241 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8777e543-c661-4a72-9221-e4d398222993-etcd-client\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.723527 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.723714 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-cqddq"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.723831 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.724352 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cqddq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.724547 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.724714 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65497b00-f979-4bad-a290-0af7be716ace-serving-cert\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.730841 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.732352 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-serving-cert\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.732535 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.734661 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mkqpg"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.738057 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.738420 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.742387 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sswrk"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.742609 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-serving-cert\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.753671 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.753964 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.754001 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.759591 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.761169 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.762388 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.763821 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566620-4lfkt"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.765375 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wd9j2"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.771087 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rfbmh"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.774761 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwrsf"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.775392 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.776005 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zdsvr"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.777131 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.778472 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-76jgq"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.785454 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qsxgq"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792130 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/71d31547-2185-4113-a21e-2f0f2ae4423c-stats-auth\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792170 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xw7gs\" (UID: \"2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792196 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8580a746-c68f-4b57-8f69-f31fb855b785-machine-approver-tls\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792217 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2-config\") pod \"kube-apiserver-operator-766d6c64bb-xw7gs\" (UID: \"2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792237 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71d31547-2185-4113-a21e-2f0f2ae4423c-service-ca-bundle\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792259 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqbg2\" (UniqueName: \"kubernetes.io/projected/45124dc2-68b5-411c-9826-be34d809a370-kube-api-access-zqbg2\") pod \"openshift-apiserver-operator-796bbdcf4f-jbsnd\" (UID: \"45124dc2-68b5-411c-9826-be34d809a370\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792283 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77917c9c-14bd-4653-ac86-571eda870a4c-serving-cert\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792302 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/77917c9c-14bd-4653-ac86-571eda870a4c-trusted-ca\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792331 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/71d31547-2185-4113-a21e-2f0f2ae4423c-metrics-certs\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792361 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bdc9b52-2233-463b-ab15-444987a74673-serving-cert\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792385 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4da55054-ed9d-4d06-a94c-5f225541f7ea-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qst64\" (UID: \"4da55054-ed9d-4d06-a94c-5f225541f7ea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792419 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwhpt\" (UniqueName: \"kubernetes.io/projected/7bdc9b52-2233-463b-ab15-444987a74673-kube-api-access-gwhpt\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792443 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bdc9b52-2233-463b-ab15-444987a74673-service-ca-bundle\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792467 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r7jj\" (UniqueName: \"kubernetes.io/projected/8580a746-c68f-4b57-8f69-f31fb855b785-kube-api-access-4r7jj\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792505 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45124dc2-68b5-411c-9826-be34d809a370-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-jbsnd\" (UID: \"45124dc2-68b5-411c-9826-be34d809a370\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792531 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77917c9c-14bd-4653-ac86-571eda870a4c-config\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792553 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jngfg\" (UniqueName: \"kubernetes.io/projected/5b3679cb-16fd-4774-88a5-15d92fe16a43-kube-api-access-jngfg\") pod \"machine-config-controller-84d6567774-sx7sz\" (UID: \"5b3679cb-16fd-4774-88a5-15d92fe16a43\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792579 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh7xz\" (UniqueName: \"kubernetes.io/projected/f49fb4d8-ad57-4153-bd98-6b21c039b971-kube-api-access-dh7xz\") pod \"openshift-config-operator-7777fb866f-2jvr2\" (UID: \"f49fb4d8-ad57-4153-bd98-6b21c039b971\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792599 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bdc9b52-2233-463b-ab15-444987a74673-config\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792624 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/249a7568-39b3-4332-a847-67a1b08d00bc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tbm9g\" (UID: \"249a7568-39b3-4332-a847-67a1b08d00bc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792647 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45124dc2-68b5-411c-9826-be34d809a370-config\") pod \"openshift-apiserver-operator-796bbdcf4f-jbsnd\" (UID: \"45124dc2-68b5-411c-9826-be34d809a370\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792668 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f49fb4d8-ad57-4153-bd98-6b21c039b971-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2jvr2\" (UID: \"f49fb4d8-ad57-4153-bd98-6b21c039b971\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792699 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bdc9b52-2233-463b-ab15-444987a74673-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792721 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/71d31547-2185-4113-a21e-2f0f2ae4423c-default-certificate\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792763 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792813 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhbvm\" (UniqueName: \"kubernetes.io/projected/71d31547-2185-4113-a21e-2f0f2ae4423c-kube-api-access-qhbvm\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792836 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f49fb4d8-ad57-4153-bd98-6b21c039b971-serving-cert\") pod \"openshift-config-operator-7777fb866f-2jvr2\" (UID: \"f49fb4d8-ad57-4153-bd98-6b21c039b971\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792857 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8580a746-c68f-4b57-8f69-f31fb855b785-config\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792879 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/249a7568-39b3-4332-a847-67a1b08d00bc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tbm9g\" (UID: \"249a7568-39b3-4332-a847-67a1b08d00bc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792913 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792935 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljfjw\" (UniqueName: \"kubernetes.io/projected/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-kube-api-access-ljfjw\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792957 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xw7gs\" (UID: \"2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.792980 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8580a746-c68f-4b57-8f69-f31fb855b785-auth-proxy-config\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.793033 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b3679cb-16fd-4774-88a5-15d92fe16a43-proxy-tls\") pod \"machine-config-controller-84d6567774-sx7sz\" (UID: \"5b3679cb-16fd-4774-88a5-15d92fe16a43\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.793177 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.793202 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/249a7568-39b3-4332-a847-67a1b08d00bc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tbm9g\" (UID: \"249a7568-39b3-4332-a847-67a1b08d00bc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.793228 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvmdp\" (UniqueName: \"kubernetes.io/projected/4da55054-ed9d-4d06-a94c-5f225541f7ea-kube-api-access-pvmdp\") pod \"control-plane-machine-set-operator-78cbb6b69f-qst64\" (UID: \"4da55054-ed9d-4d06-a94c-5f225541f7ea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.793264 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b3679cb-16fd-4774-88a5-15d92fe16a43-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sx7sz\" (UID: \"5b3679cb-16fd-4774-88a5-15d92fe16a43\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.793288 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr5hz\" (UniqueName: \"kubernetes.io/projected/77917c9c-14bd-4653-ac86-571eda870a4c-kube-api-access-dr5hz\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.795070 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45124dc2-68b5-411c-9826-be34d809a370-config\") pod \"openshift-apiserver-operator-796bbdcf4f-jbsnd\" (UID: \"45124dc2-68b5-411c-9826-be34d809a370\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.796022 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8580a746-c68f-4b57-8f69-f31fb855b785-config\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.796506 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2-config\") pod \"kube-apiserver-operator-766d6c64bb-xw7gs\" (UID: \"2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.797435 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b3679cb-16fd-4774-88a5-15d92fe16a43-proxy-tls\") pod \"machine-config-controller-84d6567774-sx7sz\" (UID: \"5b3679cb-16fd-4774-88a5-15d92fe16a43\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.797461 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/77917c9c-14bd-4653-ac86-571eda870a4c-trusted-ca\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.797523 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77917c9c-14bd-4653-ac86-571eda870a4c-config\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.797885 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qm6tg"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.798469 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b3679cb-16fd-4774-88a5-15d92fe16a43-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sx7sz\" (UID: \"5b3679cb-16fd-4774-88a5-15d92fe16a43\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.798606 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.798676 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bdc9b52-2233-463b-ab15-444987a74673-config\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.799220 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.799365 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bdc9b52-2233-463b-ab15-444987a74673-service-ca-bundle\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.799589 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bdc9b52-2233-463b-ab15-444987a74673-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.799871 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71d31547-2185-4113-a21e-2f0f2ae4423c-service-ca-bundle\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.800209 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77917c9c-14bd-4653-ac86-571eda870a4c-serving-cert\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.800444 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.800502 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8580a746-c68f-4b57-8f69-f31fb855b785-auth-proxy-config\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.800915 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/71d31547-2185-4113-a21e-2f0f2ae4423c-stats-auth\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.801109 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.801545 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/71d31547-2185-4113-a21e-2f0f2ae4423c-metrics-certs\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.801697 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f49fb4d8-ad57-4153-bd98-6b21c039b971-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2jvr2\" (UID: \"f49fb4d8-ad57-4153-bd98-6b21c039b971\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.801795 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bdc9b52-2233-463b-ab15-444987a74673-serving-cert\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.802092 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f49fb4d8-ad57-4153-bd98-6b21c039b971-serving-cert\") pod \"openshift-config-operator-7777fb866f-2jvr2\" (UID: \"f49fb4d8-ad57-4153-bd98-6b21c039b971\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.802588 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/71d31547-2185-4113-a21e-2f0f2ae4423c-default-certificate\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.803473 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45124dc2-68b5-411c-9826-be34d809a370-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-jbsnd\" (UID: \"45124dc2-68b5-411c-9826-be34d809a370\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.804183 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xw7gs\" (UID: \"2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.806640 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8580a746-c68f-4b57-8f69-f31fb855b785-machine-approver-tls\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.811505 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.813429 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.813909 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-qbx5c"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.814816 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qbx5c" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.815177 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.815983 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.816466 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.817862 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.819047 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.820141 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.821342 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.822431 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7qskq"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.823551 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hlgq9"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.824604 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.825700 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p2lfv"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.826787 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.829563 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cqddq"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.830647 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zdc26"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.830815 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4da55054-ed9d-4d06-a94c-5f225541f7ea-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qst64\" (UID: \"4da55054-ed9d-4d06-a94c-5f225541f7ea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.833884 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6cfw8"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.835219 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.835868 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.835999 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6cfw8"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.842269 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jzzk2"] Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.856063 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: W0320 09:00:47.860112 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316f49ed_2e3c_42b7_a133_a36c3074810e.slice/crio-24fb8fceda128763d89b224750a973054c51b7b2a6e0cb8d24bc586d3115859b WatchSource:0}: Error finding container 24fb8fceda128763d89b224750a973054c51b7b2a6e0cb8d24bc586d3115859b: Status 404 returned error can't find the container with id 24fb8fceda128763d89b224750a973054c51b7b2a6e0cb8d24bc586d3115859b Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.876725 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.896008 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.916658 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.936084 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.976202 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 20 09:00:47 crc kubenswrapper[4823]: I0320 09:00:47.996699 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.016403 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.029407 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/249a7568-39b3-4332-a847-67a1b08d00bc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tbm9g\" (UID: \"249a7568-39b3-4332-a847-67a1b08d00bc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.036322 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.044876 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/249a7568-39b3-4332-a847-67a1b08d00bc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tbm9g\" (UID: \"249a7568-39b3-4332-a847-67a1b08d00bc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.056364 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.076357 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.096907 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.116652 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.136294 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.156464 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.169269 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9ee12f438da2cfda54812ee1559bee5f8300a330ca6d0b3e249398764d920a9b"} Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.169327 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c0b7e42ff61d74fa3286c5e6ff54c21dbbff965254f23cfa549841b4fcec76e1"} Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.170417 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" event={"ID":"316f49ed-2e3c-42b7-a133-a36c3074810e","Type":"ContainerStarted","Data":"f1e9dbcaaa4ed7320ab23f15f0d23cfa01f2705841e14150eaa108bffc6231ce"} Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.170446 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" event={"ID":"316f49ed-2e3c-42b7-a133-a36c3074810e","Type":"ContainerStarted","Data":"24fb8fceda128763d89b224750a973054c51b7b2a6e0cb8d24bc586d3115859b"} Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.171600 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7024b4f8885c1274a704097b1394e0304a5ff406d80e4e4f65a1aba015fd76ce"} Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.171632 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e13a44bc32716bac3a2fa2323774c9e0a6506caf4a66e82d76373688188e1adf"} Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.173048 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"32421fef93ece6cb6c48a8c357114ef86f81dc69c6fa027c733f914916475a83"} Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.173113 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f427d254f6786f8a52db6e0759f556f194a83eba742b0538504ed9feb5e6f20b"} Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.173330 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.176459 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.197679 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.216514 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.236649 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.257551 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.277337 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.296562 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.316499 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.345014 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.357319 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.377521 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.416559 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.436083 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.456481 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.476965 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.496754 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.516024 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.536519 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.556461 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.576227 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.595937 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.615660 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.636187 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.656703 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.676277 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.695404 4823 request.go:700] Waited for 1.00978748s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpackage-server-manager-serving-cert&limit=500&resourceVersion=0 Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.698407 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.715771 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.737092 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.791866 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r22wq\" (UniqueName: \"kubernetes.io/projected/c8afe39d-c5e2-4241-a8ac-64f32c8c31fe-kube-api-access-r22wq\") pod \"machine-api-operator-5694c8668f-49lkt\" (UID: \"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.796978 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.803101 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtqbq\" (UniqueName: \"kubernetes.io/projected/8777e543-c661-4a72-9221-e4d398222993-kube-api-access-gtqbq\") pod \"etcd-operator-b45778765-qm6tg\" (UID: \"8777e543-c661-4a72-9221-e4d398222993\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.817039 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.837470 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.873186 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k94q\" (UniqueName: \"kubernetes.io/projected/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-kube-api-access-2k94q\") pod \"route-controller-manager-6576b87f9c-dm29j\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.902893 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g6cr\" (UniqueName: \"kubernetes.io/projected/5ce7e9b8-5c85-4ea2-ad33-14bdf64be778-kube-api-access-7g6cr\") pod \"dns-operator-744455d44c-mkqpg\" (UID: \"5ce7e9b8-5c85-4ea2-ad33-14bdf64be778\") " pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.916693 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.918652 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gczzv\" (UniqueName: \"kubernetes.io/projected/c4256007-3ebf-4991-89b5-83eb7104b8a6-kube-api-access-gczzv\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.930316 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.952668 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46s55\" (UniqueName: \"kubernetes.io/projected/e104c875-7420-45b3-b675-a63b92c454e4-kube-api-access-46s55\") pod \"console-f9d7485db-lz4n9\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.956716 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.962616 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.981682 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 20 09:00:48 crc kubenswrapper[4823]: I0320 09:00:48.998330 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.035023 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.036829 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.057952 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.073566 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.076901 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.097078 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.120603 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.130806 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.142575 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qm6tg"] Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.145045 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.153729 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.157147 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.171892 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-mkqpg"] Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.178139 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.179973 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" event={"ID":"8777e543-c661-4a72-9221-e4d398222993","Type":"ContainerStarted","Data":"470acc056b4a0216250e0914b09ffa4e47070c076a18f5e0a808dedbb834e882"} Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.183415 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jzzk2" event={"ID":"316f49ed-2e3c-42b7-a133-a36c3074810e","Type":"ContainerStarted","Data":"0a8ade8ef9f1e21df754e5e27ee3fa1dd5c44fe58a9b241eb3c7c7412a358b9e"} Mar 20 09:00:49 crc kubenswrapper[4823]: W0320 09:00:49.198227 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ce7e9b8_5c85_4ea2_ad33_14bdf64be778.slice/crio-a9afd9d01041885394ca4293562c0041cf90dec8cd3849d07ed9bc8b32aea8f6 WatchSource:0}: Error finding container a9afd9d01041885394ca4293562c0041cf90dec8cd3849d07ed9bc8b32aea8f6: Status 404 returned error can't find the container with id a9afd9d01041885394ca4293562c0041cf90dec8cd3849d07ed9bc8b32aea8f6 Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.199665 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.224914 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.254178 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zqlv\" (UniqueName: \"kubernetes.io/projected/dbf97305-8126-4377-801a-3b57bea03ea3-kube-api-access-5zqlv\") pod \"cluster-samples-operator-665b6dd947-7b46d\" (UID: \"dbf97305-8126-4377-801a-3b57bea03ea3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.259998 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-49lkt"] Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.278251 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrrcv\" (UniqueName: \"kubernetes.io/projected/98f13ee0-723b-4ef5-851b-2665009a9d47-kube-api-access-vrrcv\") pod \"downloads-7954f5f757-grjk6\" (UID: \"98f13ee0-723b-4ef5-851b-2665009a9d47\") " pod="openshift-console/downloads-7954f5f757-grjk6" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.292265 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w8b2\" (UniqueName: \"kubernetes.io/projected/65497b00-f979-4bad-a290-0af7be716ace-kube-api-access-7w8b2\") pod \"apiserver-7bbb656c7d-r28hg\" (UID: \"65497b00-f979-4bad-a290-0af7be716ace\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.298715 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.316895 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.338156 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.347599 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.357379 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.377384 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.391309 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-lz4n9"] Mar 20 09:00:49 crc kubenswrapper[4823]: W0320 09:00:49.404260 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode104c875_7420_45b3_b675_a63b92c454e4.slice/crio-006f746577c69e05f0160fea69f9a67239dcac7a2446d93eac85c8f7fcd34af2 WatchSource:0}: Error finding container 006f746577c69e05f0160fea69f9a67239dcac7a2446d93eac85c8f7fcd34af2: Status 404 returned error can't find the container with id 006f746577c69e05f0160fea69f9a67239dcac7a2446d93eac85c8f7fcd34af2 Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.415110 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c4256007-3ebf-4991-89b5-83eb7104b8a6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vn8gj\" (UID: \"c4256007-3ebf-4991-89b5-83eb7104b8a6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.435042 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vbw6\" (UniqueName: \"kubernetes.io/projected/70a71b7a-1d65-4ca4-a7c5-25a582208c0b-kube-api-access-8vbw6\") pod \"apiserver-76f77b778f-sswrk\" (UID: \"70a71b7a-1d65-4ca4-a7c5-25a582208c0b\") " pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.438904 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.454560 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j"] Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.456775 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.478244 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.479872 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.488001 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-grjk6" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.496227 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 20 09:00:49 crc kubenswrapper[4823]: W0320 09:00:49.505845 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podceb7506a_f654_4b9c_ba3d_d0fc6db798c0.slice/crio-89793889787aad51372aaaf0da12df7e5d2e7ba2309e4b4eb3418355516bb1e7 WatchSource:0}: Error finding container 89793889787aad51372aaaf0da12df7e5d2e7ba2309e4b4eb3418355516bb1e7: Status 404 returned error can't find the container with id 89793889787aad51372aaaf0da12df7e5d2e7ba2309e4b4eb3418355516bb1e7 Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.508076 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.517060 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.543252 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.556772 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.576565 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.585916 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg"] Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.601117 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 09:00:49 crc kubenswrapper[4823]: W0320 09:00:49.608547 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65497b00_f979_4bad_a290_0af7be716ace.slice/crio-565c9a6f73e3a79c26a998cc625ced394c6573c3aa778419a5036700745513f2 WatchSource:0}: Error finding container 565c9a6f73e3a79c26a998cc625ced394c6573c3aa778419a5036700745513f2: Status 404 returned error can't find the container with id 565c9a6f73e3a79c26a998cc625ced394c6573c3aa778419a5036700745513f2 Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.618393 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.649786 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.674325 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.678275 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.697253 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.697385 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.714891 4823 request.go:700] Waited for 1.990158127s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Dcanary-serving-cert&limit=500&resourceVersion=0 Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.717394 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.739720 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.763912 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-grjk6"] Mar 20 09:00:49 crc kubenswrapper[4823]: W0320 09:00:49.779144 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98f13ee0_723b_4ef5_851b_2665009a9d47.slice/crio-f00e6db5a7459a124bb89ef83da26e6a91c3fee8a691e43a06427b2ba41b025b WatchSource:0}: Error finding container f00e6db5a7459a124bb89ef83da26e6a91c3fee8a691e43a06427b2ba41b025b: Status 404 returned error can't find the container with id f00e6db5a7459a124bb89ef83da26e6a91c3fee8a691e43a06427b2ba41b025b Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.783095 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr5hz\" (UniqueName: \"kubernetes.io/projected/77917c9c-14bd-4653-ac86-571eda870a4c-kube-api-access-dr5hz\") pod \"console-operator-58897d9998-qsxgq\" (UID: \"77917c9c-14bd-4653-ac86-571eda870a4c\") " pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.791717 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xw7gs\" (UID: \"2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.812022 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d"] Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.815149 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/249a7568-39b3-4332-a847-67a1b08d00bc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tbm9g\" (UID: \"249a7568-39b3-4332-a847-67a1b08d00bc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.836106 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.854274 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhbvm\" (UniqueName: \"kubernetes.io/projected/71d31547-2185-4113-a21e-2f0f2ae4423c-kube-api-access-qhbvm\") pod \"router-default-5444994796-lndmn\" (UID: \"71d31547-2185-4113-a21e-2f0f2ae4423c\") " pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.871790 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvmdp\" (UniqueName: \"kubernetes.io/projected/4da55054-ed9d-4d06-a94c-5f225541f7ea-kube-api-access-pvmdp\") pod \"control-plane-machine-set-operator-78cbb6b69f-qst64\" (UID: \"4da55054-ed9d-4d06-a94c-5f225541f7ea\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.887931 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.895158 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.895704 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh7xz\" (UniqueName: \"kubernetes.io/projected/f49fb4d8-ad57-4153-bd98-6b21c039b971-kube-api-access-dh7xz\") pod \"openshift-config-operator-7777fb866f-2jvr2\" (UID: \"f49fb4d8-ad57-4153-bd98-6b21c039b971\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.899219 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.916359 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj"] Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.918311 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.920089 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jngfg\" (UniqueName: \"kubernetes.io/projected/5b3679cb-16fd-4774-88a5-15d92fe16a43-kube-api-access-jngfg\") pod \"machine-config-controller-84d6567774-sx7sz\" (UID: \"5b3679cb-16fd-4774-88a5-15d92fe16a43\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.927820 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sswrk"] Mar 20 09:00:49 crc kubenswrapper[4823]: W0320 09:00:49.935463 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4256007_3ebf_4991_89b5_83eb7104b8a6.slice/crio-64303fea76567f80ec24e4f75e1a6541a24437d524eb0579de67424a19d42f98 WatchSource:0}: Error finding container 64303fea76567f80ec24e4f75e1a6541a24437d524eb0579de67424a19d42f98: Status 404 returned error can't find the container with id 64303fea76567f80ec24e4f75e1a6541a24437d524eb0579de67424a19d42f98 Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.935733 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljfjw\" (UniqueName: \"kubernetes.io/projected/e64e0f18-5f66-49ca-9cca-b97bde4f2dc6-kube-api-access-ljfjw\") pod \"cluster-image-registry-operator-dc59b4c8b-pcdfn\" (UID: \"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.937565 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" Mar 20 09:00:49 crc kubenswrapper[4823]: W0320 09:00:49.944458 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70a71b7a_1d65_4ca4_a7c5_25a582208c0b.slice/crio-0bd64066bf969a9b8cab0cf5c3ba26e4d4ba296aefbb422b0922e345835ea137 WatchSource:0}: Error finding container 0bd64066bf969a9b8cab0cf5c3ba26e4d4ba296aefbb422b0922e345835ea137: Status 404 returned error can't find the container with id 0bd64066bf969a9b8cab0cf5c3ba26e4d4ba296aefbb422b0922e345835ea137 Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.953160 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwhpt\" (UniqueName: \"kubernetes.io/projected/7bdc9b52-2233-463b-ab15-444987a74673-kube-api-access-gwhpt\") pod \"authentication-operator-69f744f599-zdsvr\" (UID: \"7bdc9b52-2233-463b-ab15-444987a74673\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.973344 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqbg2\" (UniqueName: \"kubernetes.io/projected/45124dc2-68b5-411c-9826-be34d809a370-kube-api-access-zqbg2\") pod \"openshift-apiserver-operator-796bbdcf4f-jbsnd\" (UID: \"45124dc2-68b5-411c-9826-be34d809a370\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.994219 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r7jj\" (UniqueName: \"kubernetes.io/projected/8580a746-c68f-4b57-8f69-f31fb855b785-kube-api-access-4r7jj\") pod \"machine-approver-56656f9798-gkv7j\" (UID: \"8580a746-c68f-4b57-8f69-f31fb855b785\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:49 crc kubenswrapper[4823]: I0320 09:00:49.996097 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.017002 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.039910 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.048316 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.059074 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.077619 4823 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.078996 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.088621 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.097788 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157194 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh4pq\" (UniqueName: \"kubernetes.io/projected/7d2b41be-9ca2-4c6e-a643-23265bdb2ec8-kube-api-access-hh4pq\") pod \"openshift-controller-manager-operator-756b6f6bc6-7bbkt\" (UID: \"7d2b41be-9ca2-4c6e-a643-23265bdb2ec8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157266 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-bound-sa-token\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157421 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbrcl\" (UniqueName: \"kubernetes.io/projected/e17f16e4-ac2a-417a-a657-edb90df6622c-kube-api-access-sbrcl\") pod \"kube-storage-version-migrator-operator-b67b599dd-p656g\" (UID: \"e17f16e4-ac2a-417a-a657-edb90df6622c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157441 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-certificates\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157467 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a391e636-89df-41fd-9148-9500306e1e0c-config\") pod \"kube-controller-manager-operator-78b949d7b-r4kcw\" (UID: \"a391e636-89df-41fd-9148-9500306e1e0c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157504 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e9000667-ed2e-4fea-b03f-c9760e59a604-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157520 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a391e636-89df-41fd-9148-9500306e1e0c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r4kcw\" (UID: \"a391e636-89df-41fd-9148-9500306e1e0c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157573 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6nnp\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-kube-api-access-n6nnp\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157589 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e17f16e4-ac2a-417a-a657-edb90df6622c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-p656g\" (UID: \"e17f16e4-ac2a-417a-a657-edb90df6622c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157621 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wd9j2\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157677 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e17f16e4-ac2a-417a-a657-edb90df6622c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-p656g\" (UID: \"e17f16e4-ac2a-417a-a657-edb90df6622c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157694 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q29tf\" (UniqueName: \"kubernetes.io/projected/b54b2f8f-1e5a-44cb-ac82-77040375eb02-kube-api-access-q29tf\") pod \"migrator-59844c95c7-t8r2x\" (UID: \"b54b2f8f-1e5a-44cb-ac82-77040375eb02\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157709 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d2b41be-9ca2-4c6e-a643-23265bdb2ec8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7bbkt\" (UID: \"7d2b41be-9ca2-4c6e-a643-23265bdb2ec8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157777 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wd9j2\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157826 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e9000667-ed2e-4fea-b03f-c9760e59a604-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157841 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-trusted-ca\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157874 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157900 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a391e636-89df-41fd-9148-9500306e1e0c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r4kcw\" (UID: \"a391e636-89df-41fd-9148-9500306e1e0c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157923 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fxlf\" (UniqueName: \"kubernetes.io/projected/d31847cb-ea3f-46b1-ad2a-12d87b23d882-kube-api-access-6fxlf\") pod \"marketplace-operator-79b997595-wd9j2\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157955 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-tls\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.157971 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d2b41be-9ca2-4c6e-a643-23265bdb2ec8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7bbkt\" (UID: \"7d2b41be-9ca2-4c6e-a643-23265bdb2ec8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" Mar 20 09:00:50 crc kubenswrapper[4823]: E0320 09:00:50.161342 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:50.661327053 +0000 UTC m=+217.347223427 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.177140 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.185871 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.205939 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.222116 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" event={"ID":"5ce7e9b8-5c85-4ea2-ad33-14bdf64be778","Type":"ContainerStarted","Data":"4b142e8e934df10b2eac3fe3f02db6df85ee9f26d20732d19f13254372b99eef"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.222161 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" event={"ID":"5ce7e9b8-5c85-4ea2-ad33-14bdf64be778","Type":"ContainerStarted","Data":"20718e2e91a7f4c5faf014a2e1e56783836f50f2ad297bc6335bdf402eb2dbe3"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.222170 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" event={"ID":"5ce7e9b8-5c85-4ea2-ad33-14bdf64be778","Type":"ContainerStarted","Data":"a9afd9d01041885394ca4293562c0041cf90dec8cd3849d07ed9bc8b32aea8f6"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259139 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259342 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt2wq\" (UniqueName: \"kubernetes.io/projected/c4aad9d5-b289-45b4-974c-b83bcf1b900f-kube-api-access-xt2wq\") pod \"service-ca-9c57cc56f-zdc26\" (UID: \"c4aad9d5-b289-45b4-974c-b83bcf1b900f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259393 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-certificates\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259419 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a30d4c31-5122-48fa-8af7-1eba7d2f4254-certs\") pod \"machine-config-server-qbx5c\" (UID: \"a30d4c31-5122-48fa-8af7-1eba7d2f4254\") " pod="openshift-machine-config-operator/machine-config-server-qbx5c" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259442 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259477 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a30d4c31-5122-48fa-8af7-1eba7d2f4254-node-bootstrap-token\") pod \"machine-config-server-qbx5c\" (UID: \"a30d4c31-5122-48fa-8af7-1eba7d2f4254\") " pod="openshift-machine-config-operator/machine-config-server-qbx5c" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259518 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66a777fc-e931-46d8-bf30-a0ab7f334181-images\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259564 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a391e636-89df-41fd-9148-9500306e1e0c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r4kcw\" (UID: \"a391e636-89df-41fd-9148-9500306e1e0c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259586 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259608 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6e5cca93-e5bb-4297-bf2d-88af3ca984d2-profile-collector-cert\") pod \"catalog-operator-68c6474976-jjkbd\" (UID: \"6e5cca93-e5bb-4297-bf2d-88af3ca984d2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259634 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/66a777fc-e931-46d8-bf30-a0ab7f334181-auth-proxy-config\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259672 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1d45028a-7db5-4071-901d-2c392d6730b6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g2xtb\" (UID: \"1d45028a-7db5-4071-901d-2c392d6730b6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259699 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e17f16e4-ac2a-417a-a657-edb90df6622c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-p656g\" (UID: \"e17f16e4-ac2a-417a-a657-edb90df6622c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259725 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66cb5542-0914-40bf-b27d-115a56b07012-audit-dir\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259763 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259788 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-socket-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259838 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wd9j2\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259878 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncldg\" (UniqueName: \"kubernetes.io/projected/65a02b02-7f51-4774-b050-292c3fd4f70b-kube-api-access-ncldg\") pod \"auto-csr-approver-29566620-4lfkt\" (UID: \"65a02b02-7f51-4774-b050-292c3fd4f70b\") " pod="openshift-infra/auto-csr-approver-29566620-4lfkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259915 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e17f16e4-ac2a-417a-a657-edb90df6622c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-p656g\" (UID: \"e17f16e4-ac2a-417a-a657-edb90df6622c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259951 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57tcj\" (UniqueName: \"kubernetes.io/projected/3e566079-aa6d-4037-a58b-efc6721bc079-kube-api-access-57tcj\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.259997 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0ccc792-7011-435d-934e-ad2d9d94ab90-serving-cert\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260021 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-plugins-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260041 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/64de3631-c237-4cb6-9b80-aca29c50d694-tmpfs\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260061 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfvq2\" (UniqueName: \"kubernetes.io/projected/64de3631-c237-4cb6-9b80-aca29c50d694-kube-api-access-rfvq2\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260099 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260134 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9pvw\" (UniqueName: \"kubernetes.io/projected/b1d558f6-5989-4d9e-8cfb-d5bd47b42948-kube-api-access-p9pvw\") pod \"multus-admission-controller-857f4d67dd-rfbmh\" (UID: \"b1d558f6-5989-4d9e-8cfb-d5bd47b42948\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260190 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-registration-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260215 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl9dg\" (UniqueName: \"kubernetes.io/projected/e0ccc792-7011-435d-934e-ad2d9d94ab90-kube-api-access-jl9dg\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260236 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-secret-volume\") pod \"collect-profiles-29566620-smfx8\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260259 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260293 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1b7e18eb-6077-432c-a3b0-383a51b9857d-srv-cert\") pod \"olm-operator-6b444d44fb-9r5tj\" (UID: \"1b7e18eb-6077-432c-a3b0-383a51b9857d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260312 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6e5cca93-e5bb-4297-bf2d-88af3ca984d2-srv-cert\") pod \"catalog-operator-68c6474976-jjkbd\" (UID: \"6e5cca93-e5bb-4297-bf2d-88af3ca984d2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260374 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/362283f3-bd9a-4c9e-8dbb-b4b3358ed5de-config\") pod \"service-ca-operator-777779d784-7qskq\" (UID: \"362283f3-bd9a-4c9e-8dbb-b4b3358ed5de\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260400 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a391e636-89df-41fd-9148-9500306e1e0c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r4kcw\" (UID: \"a391e636-89df-41fd-9148-9500306e1e0c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260423 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9f8g\" (UniqueName: \"kubernetes.io/projected/1a8b9996-59a1-4167-8943-2c8e66cd1d3c-kube-api-access-w9f8g\") pod \"dns-default-76jgq\" (UID: \"1a8b9996-59a1-4167-8943-2c8e66cd1d3c\") " pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260445 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-csi-data-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260466 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260522 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a8b9996-59a1-4167-8943-2c8e66cd1d3c-config-volume\") pod \"dns-default-76jgq\" (UID: \"1a8b9996-59a1-4167-8943-2c8e66cd1d3c\") " pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260543 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/64de3631-c237-4cb6-9b80-aca29c50d694-apiservice-cert\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260587 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6d934348-d496-4ffa-8220-4c7d592a59ce-cert\") pod \"ingress-canary-cqddq\" (UID: \"6d934348-d496-4ffa-8220-4c7d592a59ce\") " pod="openshift-ingress-canary/ingress-canary-cqddq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260621 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-bound-sa-token\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260646 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh4pq\" (UniqueName: \"kubernetes.io/projected/7d2b41be-9ca2-4c6e-a643-23265bdb2ec8-kube-api-access-hh4pq\") pod \"openshift-controller-manager-operator-756b6f6bc6-7bbkt\" (UID: \"7d2b41be-9ca2-4c6e-a643-23265bdb2ec8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260667 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a8b9996-59a1-4167-8943-2c8e66cd1d3c-metrics-tls\") pod \"dns-default-76jgq\" (UID: \"1a8b9996-59a1-4167-8943-2c8e66cd1d3c\") " pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260689 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbrcl\" (UniqueName: \"kubernetes.io/projected/e17f16e4-ac2a-417a-a657-edb90df6622c-kube-api-access-sbrcl\") pod \"kube-storage-version-migrator-operator-b67b599dd-p656g\" (UID: \"e17f16e4-ac2a-417a-a657-edb90df6622c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260752 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a391e636-89df-41fd-9148-9500306e1e0c-config\") pod \"kube-controller-manager-operator-78b949d7b-r4kcw\" (UID: \"a391e636-89df-41fd-9148-9500306e1e0c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260821 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/362283f3-bd9a-4c9e-8dbb-b4b3358ed5de-serving-cert\") pod \"service-ca-operator-777779d784-7qskq\" (UID: \"362283f3-bd9a-4c9e-8dbb-b4b3358ed5de\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260849 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260875 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c4aad9d5-b289-45b4-974c-b83bcf1b900f-signing-cabundle\") pod \"service-ca-9c57cc56f-zdc26\" (UID: \"c4aad9d5-b289-45b4-974c-b83bcf1b900f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260897 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e9000667-ed2e-4fea-b03f-c9760e59a604-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260919 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gkgx\" (UniqueName: \"kubernetes.io/projected/a30d4c31-5122-48fa-8af7-1eba7d2f4254-kube-api-access-9gkgx\") pod \"machine-config-server-qbx5c\" (UID: \"a30d4c31-5122-48fa-8af7-1eba7d2f4254\") " pod="openshift-machine-config-operator/machine-config-server-qbx5c" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260943 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qfjh\" (UniqueName: \"kubernetes.io/projected/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-kube-api-access-9qfjh\") pod \"collect-profiles-29566620-smfx8\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260972 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvqg5\" (UniqueName: \"kubernetes.io/projected/6e5cca93-e5bb-4297-bf2d-88af3ca984d2-kube-api-access-qvqg5\") pod \"catalog-operator-68c6474976-jjkbd\" (UID: \"6e5cca93-e5bb-4297-bf2d-88af3ca984d2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.260994 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.261029 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-audit-policies\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.261092 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6nnp\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-kube-api-access-n6nnp\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.261130 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1b7e18eb-6077-432c-a3b0-383a51b9857d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9r5tj\" (UID: \"1b7e18eb-6077-432c-a3b0-383a51b9857d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.261151 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhv6m\" (UniqueName: \"kubernetes.io/projected/1b7e18eb-6077-432c-a3b0-383a51b9857d-kube-api-access-jhv6m\") pod \"olm-operator-6b444d44fb-9r5tj\" (UID: \"1b7e18eb-6077-432c-a3b0-383a51b9857d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.261186 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.261881 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q29tf\" (UniqueName: \"kubernetes.io/projected/b54b2f8f-1e5a-44cb-ac82-77040375eb02-kube-api-access-q29tf\") pod \"migrator-59844c95c7-t8r2x\" (UID: \"b54b2f8f-1e5a-44cb-ac82-77040375eb02\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.261923 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d2b41be-9ca2-4c6e-a643-23265bdb2ec8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7bbkt\" (UID: \"7d2b41be-9ca2-4c6e-a643-23265bdb2ec8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.261948 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/66a777fc-e931-46d8-bf30-a0ab7f334181-proxy-tls\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262006 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wd9j2\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262058 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d558f6-5989-4d9e-8cfb-d5bd47b42948-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rfbmh\" (UID: \"b1d558f6-5989-4d9e-8cfb-d5bd47b42948\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262117 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmphk\" (UniqueName: \"kubernetes.io/projected/66a777fc-e931-46d8-bf30-a0ab7f334181-kube-api-access-lmphk\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262142 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e9000667-ed2e-4fea-b03f-c9760e59a604-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262164 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-trusted-ca\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262204 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262226 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-mountpoint-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262247 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-client-ca\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262284 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2r2b\" (UniqueName: \"kubernetes.io/projected/66cb5542-0914-40bf-b27d-115a56b07012-kube-api-access-b2r2b\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262307 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w85l6\" (UniqueName: \"kubernetes.io/projected/362283f3-bd9a-4c9e-8dbb-b4b3358ed5de-kube-api-access-w85l6\") pod \"service-ca-operator-777779d784-7qskq\" (UID: \"362283f3-bd9a-4c9e-8dbb-b4b3358ed5de\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262330 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/64de3631-c237-4cb6-9b80-aca29c50d694-webhook-cert\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262369 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k2hn\" (UniqueName: \"kubernetes.io/projected/1d45028a-7db5-4071-901d-2c392d6730b6-kube-api-access-5k2hn\") pod \"package-server-manager-789f6589d5-g2xtb\" (UID: \"1d45028a-7db5-4071-901d-2c392d6730b6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262390 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c4aad9d5-b289-45b4-974c-b83bcf1b900f-signing-key\") pod \"service-ca-9c57cc56f-zdc26\" (UID: \"c4aad9d5-b289-45b4-974c-b83bcf1b900f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262412 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262442 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-tls\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262481 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fxlf\" (UniqueName: \"kubernetes.io/projected/d31847cb-ea3f-46b1-ad2a-12d87b23d882-kube-api-access-6fxlf\") pod \"marketplace-operator-79b997595-wd9j2\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262509 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d2b41be-9ca2-4c6e-a643-23265bdb2ec8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7bbkt\" (UID: \"7d2b41be-9ca2-4c6e-a643-23265bdb2ec8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262538 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-config\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262565 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt265\" (UniqueName: \"kubernetes.io/projected/6d934348-d496-4ffa-8220-4c7d592a59ce-kube-api-access-dt265\") pod \"ingress-canary-cqddq\" (UID: \"6d934348-d496-4ffa-8220-4c7d592a59ce\") " pod="openshift-ingress-canary/ingress-canary-cqddq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262590 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-config-volume\") pod \"collect-profiles-29566620-smfx8\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.262633 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: E0320 09:00:50.263623 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:50.76359884 +0000 UTC m=+217.449495214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.267921 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-certificates\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.272004 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a391e636-89df-41fd-9148-9500306e1e0c-config\") pod \"kube-controller-manager-operator-78b949d7b-r4kcw\" (UID: \"a391e636-89df-41fd-9148-9500306e1e0c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.279435 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d2b41be-9ca2-4c6e-a643-23265bdb2ec8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7bbkt\" (UID: \"7d2b41be-9ca2-4c6e-a643-23265bdb2ec8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.300274 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a391e636-89df-41fd-9148-9500306e1e0c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-r4kcw\" (UID: \"a391e636-89df-41fd-9148-9500306e1e0c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.300642 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wd9j2\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.302214 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e9000667-ed2e-4fea-b03f-c9760e59a604-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.302443 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e17f16e4-ac2a-417a-a657-edb90df6622c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-p656g\" (UID: \"e17f16e4-ac2a-417a-a657-edb90df6622c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.302614 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e17f16e4-ac2a-417a-a657-edb90df6622c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-p656g\" (UID: \"e17f16e4-ac2a-417a-a657-edb90df6622c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.304553 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64"] Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.306651 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-grjk6" event={"ID":"98f13ee0-723b-4ef5-851b-2665009a9d47","Type":"ContainerStarted","Data":"7abd020b6d7b4d5b68f635edb30b364184211e34dad972060b73d25b0dfcc926"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.306709 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-grjk6" event={"ID":"98f13ee0-723b-4ef5-851b-2665009a9d47","Type":"ContainerStarted","Data":"f00e6db5a7459a124bb89ef83da26e6a91c3fee8a691e43a06427b2ba41b025b"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.306994 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e9000667-ed2e-4fea-b03f-c9760e59a604-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.307236 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-grjk6" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.308883 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d2b41be-9ca2-4c6e-a643-23265bdb2ec8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7bbkt\" (UID: \"7d2b41be-9ca2-4c6e-a643-23265bdb2ec8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.314457 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs"] Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.315663 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wd9j2\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.318980 4823 patch_prober.go:28] interesting pod/downloads-7954f5f757-grjk6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.319082 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grjk6" podUID="98f13ee0-723b-4ef5-851b-2665009a9d47" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.320026 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-tls\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.332242 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-trusted-ca\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.333095 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" event={"ID":"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe","Type":"ContainerStarted","Data":"fb7dc0f6d7b21ad25766428b11b5bc5a57bfe92941d0ad95890cf06c6b903cd5"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.333136 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" event={"ID":"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe","Type":"ContainerStarted","Data":"1ee5d132c9eafe3fab36b4cb4d928eb3d97f77e12df79ce77ceb6b8d4fbf8d50"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.333147 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" event={"ID":"c8afe39d-c5e2-4241-a8ac-64f32c8c31fe","Type":"ContainerStarted","Data":"3c00e8117351be01e36edc9fe1fad810e9c8feba8b646d977a21535d680fdc3e"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.337140 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a391e636-89df-41fd-9148-9500306e1e0c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-r4kcw\" (UID: \"a391e636-89df-41fd-9148-9500306e1e0c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.339280 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6nnp\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-kube-api-access-n6nnp\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363657 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66a777fc-e931-46d8-bf30-a0ab7f334181-images\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363706 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363726 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/66a777fc-e931-46d8-bf30-a0ab7f334181-auth-proxy-config\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363760 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1d45028a-7db5-4071-901d-2c392d6730b6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g2xtb\" (UID: \"1d45028a-7db5-4071-901d-2c392d6730b6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363778 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6e5cca93-e5bb-4297-bf2d-88af3ca984d2-profile-collector-cert\") pod \"catalog-operator-68c6474976-jjkbd\" (UID: \"6e5cca93-e5bb-4297-bf2d-88af3ca984d2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363797 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66cb5542-0914-40bf-b27d-115a56b07012-audit-dir\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363814 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363829 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-socket-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363849 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncldg\" (UniqueName: \"kubernetes.io/projected/65a02b02-7f51-4774-b050-292c3fd4f70b-kube-api-access-ncldg\") pod \"auto-csr-approver-29566620-4lfkt\" (UID: \"65a02b02-7f51-4774-b050-292c3fd4f70b\") " pod="openshift-infra/auto-csr-approver-29566620-4lfkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363869 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57tcj\" (UniqueName: \"kubernetes.io/projected/3e566079-aa6d-4037-a58b-efc6721bc079-kube-api-access-57tcj\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363884 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0ccc792-7011-435d-934e-ad2d9d94ab90-serving-cert\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363898 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-plugins-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363915 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/64de3631-c237-4cb6-9b80-aca29c50d694-tmpfs\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363971 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfvq2\" (UniqueName: \"kubernetes.io/projected/64de3631-c237-4cb6-9b80-aca29c50d694-kube-api-access-rfvq2\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.363990 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364010 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9pvw\" (UniqueName: \"kubernetes.io/projected/b1d558f6-5989-4d9e-8cfb-d5bd47b42948-kube-api-access-p9pvw\") pod \"multus-admission-controller-857f4d67dd-rfbmh\" (UID: \"b1d558f6-5989-4d9e-8cfb-d5bd47b42948\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364029 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-registration-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364052 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl9dg\" (UniqueName: \"kubernetes.io/projected/e0ccc792-7011-435d-934e-ad2d9d94ab90-kube-api-access-jl9dg\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364074 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-secret-volume\") pod \"collect-profiles-29566620-smfx8\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364100 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364115 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1b7e18eb-6077-432c-a3b0-383a51b9857d-srv-cert\") pod \"olm-operator-6b444d44fb-9r5tj\" (UID: \"1b7e18eb-6077-432c-a3b0-383a51b9857d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364134 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6e5cca93-e5bb-4297-bf2d-88af3ca984d2-srv-cert\") pod \"catalog-operator-68c6474976-jjkbd\" (UID: \"6e5cca93-e5bb-4297-bf2d-88af3ca984d2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364159 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/362283f3-bd9a-4c9e-8dbb-b4b3358ed5de-config\") pod \"service-ca-operator-777779d784-7qskq\" (UID: \"362283f3-bd9a-4c9e-8dbb-b4b3358ed5de\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364182 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9f8g\" (UniqueName: \"kubernetes.io/projected/1a8b9996-59a1-4167-8943-2c8e66cd1d3c-kube-api-access-w9f8g\") pod \"dns-default-76jgq\" (UID: \"1a8b9996-59a1-4167-8943-2c8e66cd1d3c\") " pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364219 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-csi-data-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364256 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a8b9996-59a1-4167-8943-2c8e66cd1d3c-config-volume\") pod \"dns-default-76jgq\" (UID: \"1a8b9996-59a1-4167-8943-2c8e66cd1d3c\") " pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364279 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364301 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/64de3631-c237-4cb6-9b80-aca29c50d694-apiservice-cert\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364325 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6d934348-d496-4ffa-8220-4c7d592a59ce-cert\") pod \"ingress-canary-cqddq\" (UID: \"6d934348-d496-4ffa-8220-4c7d592a59ce\") " pod="openshift-ingress-canary/ingress-canary-cqddq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364362 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a8b9996-59a1-4167-8943-2c8e66cd1d3c-metrics-tls\") pod \"dns-default-76jgq\" (UID: \"1a8b9996-59a1-4167-8943-2c8e66cd1d3c\") " pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364397 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/362283f3-bd9a-4c9e-8dbb-b4b3358ed5de-serving-cert\") pod \"service-ca-operator-777779d784-7qskq\" (UID: \"362283f3-bd9a-4c9e-8dbb-b4b3358ed5de\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364421 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364444 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c4aad9d5-b289-45b4-974c-b83bcf1b900f-signing-cabundle\") pod \"service-ca-9c57cc56f-zdc26\" (UID: \"c4aad9d5-b289-45b4-974c-b83bcf1b900f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364482 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gkgx\" (UniqueName: \"kubernetes.io/projected/a30d4c31-5122-48fa-8af7-1eba7d2f4254-kube-api-access-9gkgx\") pod \"machine-config-server-qbx5c\" (UID: \"a30d4c31-5122-48fa-8af7-1eba7d2f4254\") " pod="openshift-machine-config-operator/machine-config-server-qbx5c" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364507 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qfjh\" (UniqueName: \"kubernetes.io/projected/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-kube-api-access-9qfjh\") pod \"collect-profiles-29566620-smfx8\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364529 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvqg5\" (UniqueName: \"kubernetes.io/projected/6e5cca93-e5bb-4297-bf2d-88af3ca984d2-kube-api-access-qvqg5\") pod \"catalog-operator-68c6474976-jjkbd\" (UID: \"6e5cca93-e5bb-4297-bf2d-88af3ca984d2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364553 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364576 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-audit-policies\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364603 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1b7e18eb-6077-432c-a3b0-383a51b9857d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9r5tj\" (UID: \"1b7e18eb-6077-432c-a3b0-383a51b9857d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364628 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhv6m\" (UniqueName: \"kubernetes.io/projected/1b7e18eb-6077-432c-a3b0-383a51b9857d-kube-api-access-jhv6m\") pod \"olm-operator-6b444d44fb-9r5tj\" (UID: \"1b7e18eb-6077-432c-a3b0-383a51b9857d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364652 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364699 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/66a777fc-e931-46d8-bf30-a0ab7f334181-proxy-tls\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364727 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d558f6-5989-4d9e-8cfb-d5bd47b42948-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rfbmh\" (UID: \"b1d558f6-5989-4d9e-8cfb-d5bd47b42948\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364795 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmphk\" (UniqueName: \"kubernetes.io/projected/66a777fc-e931-46d8-bf30-a0ab7f334181-kube-api-access-lmphk\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364825 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364849 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-mountpoint-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364872 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-client-ca\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364902 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364931 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2r2b\" (UniqueName: \"kubernetes.io/projected/66cb5542-0914-40bf-b27d-115a56b07012-kube-api-access-b2r2b\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364955 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w85l6\" (UniqueName: \"kubernetes.io/projected/362283f3-bd9a-4c9e-8dbb-b4b3358ed5de-kube-api-access-w85l6\") pod \"service-ca-operator-777779d784-7qskq\" (UID: \"362283f3-bd9a-4c9e-8dbb-b4b3358ed5de\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.364977 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/64de3631-c237-4cb6-9b80-aca29c50d694-webhook-cert\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365000 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k2hn\" (UniqueName: \"kubernetes.io/projected/1d45028a-7db5-4071-901d-2c392d6730b6-kube-api-access-5k2hn\") pod \"package-server-manager-789f6589d5-g2xtb\" (UID: \"1d45028a-7db5-4071-901d-2c392d6730b6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365024 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365057 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c4aad9d5-b289-45b4-974c-b83bcf1b900f-signing-key\") pod \"service-ca-9c57cc56f-zdc26\" (UID: \"c4aad9d5-b289-45b4-974c-b83bcf1b900f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365080 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-config\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365102 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt265\" (UniqueName: \"kubernetes.io/projected/6d934348-d496-4ffa-8220-4c7d592a59ce-kube-api-access-dt265\") pod \"ingress-canary-cqddq\" (UID: \"6d934348-d496-4ffa-8220-4c7d592a59ce\") " pod="openshift-ingress-canary/ingress-canary-cqddq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365129 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-config-volume\") pod \"collect-profiles-29566620-smfx8\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365159 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365182 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt2wq\" (UniqueName: \"kubernetes.io/projected/c4aad9d5-b289-45b4-974c-b83bcf1b900f-kube-api-access-xt2wq\") pod \"service-ca-9c57cc56f-zdc26\" (UID: \"c4aad9d5-b289-45b4-974c-b83bcf1b900f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365205 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a30d4c31-5122-48fa-8af7-1eba7d2f4254-certs\") pod \"machine-config-server-qbx5c\" (UID: \"a30d4c31-5122-48fa-8af7-1eba7d2f4254\") " pod="openshift-machine-config-operator/machine-config-server-qbx5c" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365227 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365254 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a30d4c31-5122-48fa-8af7-1eba7d2f4254-node-bootstrap-token\") pod \"machine-config-server-qbx5c\" (UID: \"a30d4c31-5122-48fa-8af7-1eba7d2f4254\") " pod="openshift-machine-config-operator/machine-config-server-qbx5c" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.365292 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-bound-sa-token\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.366492 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-audit-policies\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.366573 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.370917 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.371699 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c4aad9d5-b289-45b4-974c-b83bcf1b900f-signing-cabundle\") pod \"service-ca-9c57cc56f-zdc26\" (UID: \"c4aad9d5-b289-45b4-974c-b83bcf1b900f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.371970 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" event={"ID":"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0","Type":"ContainerStarted","Data":"996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.372010 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" event={"ID":"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0","Type":"ContainerStarted","Data":"89793889787aad51372aaaf0da12df7e5d2e7ba2309e4b4eb3418355516bb1e7"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.377029 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.380061 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66a777fc-e931-46d8-bf30-a0ab7f334181-images\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.383022 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-registration-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.383481 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-csi-data-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.383755 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.383790 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-mountpoint-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.384537 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-client-ca\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: E0320 09:00:50.384831 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:50.8848168 +0000 UTC m=+217.570713174 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.390917 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q29tf\" (UniqueName: \"kubernetes.io/projected/b54b2f8f-1e5a-44cb-ac82-77040375eb02-kube-api-access-q29tf\") pod \"migrator-59844c95c7-t8r2x\" (UID: \"b54b2f8f-1e5a-44cb-ac82-77040375eb02\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.391872 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1b7e18eb-6077-432c-a3b0-383a51b9857d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9r5tj\" (UID: \"1b7e18eb-6077-432c-a3b0-383a51b9857d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.392180 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a30d4c31-5122-48fa-8af7-1eba7d2f4254-node-bootstrap-token\") pod \"machine-config-server-qbx5c\" (UID: \"a30d4c31-5122-48fa-8af7-1eba7d2f4254\") " pod="openshift-machine-config-operator/machine-config-server-qbx5c" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.396172 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/66a777fc-e931-46d8-bf30-a0ab7f334181-auth-proxy-config\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.396391 4823 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dm29j container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.396452 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" podUID="ceb7506a-f654-4b9c-ba3d-d0fc6db798c0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.397368 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g"] Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.400786 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/362283f3-bd9a-4c9e-8dbb-b4b3358ed5de-serving-cert\") pod \"service-ca-operator-777779d784-7qskq\" (UID: \"362283f3-bd9a-4c9e-8dbb-b4b3358ed5de\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.404719 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66cb5542-0914-40bf-b27d-115a56b07012-audit-dir\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.406143 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1d45028a-7db5-4071-901d-2c392d6730b6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g2xtb\" (UID: \"1d45028a-7db5-4071-901d-2c392d6730b6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.406553 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-socket-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.410668 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-config\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.411460 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-config-volume\") pod \"collect-profiles-29566620-smfx8\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.414811 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/64de3631-c237-4cb6-9b80-aca29c50d694-apiservice-cert\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.417405 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6e5cca93-e5bb-4297-bf2d-88af3ca984d2-profile-collector-cert\") pod \"catalog-operator-68c6474976-jjkbd\" (UID: \"6e5cca93-e5bb-4297-bf2d-88af3ca984d2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.417617 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c4aad9d5-b289-45b4-974c-b83bcf1b900f-signing-key\") pod \"service-ca-9c57cc56f-zdc26\" (UID: \"c4aad9d5-b289-45b4-974c-b83bcf1b900f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.391873 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a8b9996-59a1-4167-8943-2c8e66cd1d3c-config-volume\") pod \"dns-default-76jgq\" (UID: \"1a8b9996-59a1-4167-8943-2c8e66cd1d3c\") " pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.419046 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.419080 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sswrk" event={"ID":"70a71b7a-1d65-4ca4-a7c5-25a582208c0b","Type":"ContainerStarted","Data":"0bd64066bf969a9b8cab0cf5c3ba26e4d4ba296aefbb422b0922e345835ea137"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.419131 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.419366 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3e566079-aa6d-4037-a58b-efc6721bc079-plugins-dir\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.419463 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/64de3631-c237-4cb6-9b80-aca29c50d694-tmpfs\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.419716 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/362283f3-bd9a-4c9e-8dbb-b4b3358ed5de-config\") pod \"service-ca-operator-777779d784-7qskq\" (UID: \"362283f3-bd9a-4c9e-8dbb-b4b3358ed5de\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.420099 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a8b9996-59a1-4167-8943-2c8e66cd1d3c-metrics-tls\") pod \"dns-default-76jgq\" (UID: \"1a8b9996-59a1-4167-8943-2c8e66cd1d3c\") " pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.421895 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.423081 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d558f6-5989-4d9e-8cfb-d5bd47b42948-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rfbmh\" (UID: \"b1d558f6-5989-4d9e-8cfb-d5bd47b42948\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.427626 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-secret-volume\") pod \"collect-profiles-29566620-smfx8\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.431593 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1b7e18eb-6077-432c-a3b0-383a51b9857d-srv-cert\") pod \"olm-operator-6b444d44fb-9r5tj\" (UID: \"1b7e18eb-6077-432c-a3b0-383a51b9857d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.432191 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6d934348-d496-4ffa-8220-4c7d592a59ce-cert\") pod \"ingress-canary-cqddq\" (UID: \"6d934348-d496-4ffa-8220-4c7d592a59ce\") " pod="openshift-ingress-canary/ingress-canary-cqddq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.432362 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.432765 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6e5cca93-e5bb-4297-bf2d-88af3ca984d2-srv-cert\") pod \"catalog-operator-68c6474976-jjkbd\" (UID: \"6e5cca93-e5bb-4297-bf2d-88af3ca984d2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.433050 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.433796 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.433955 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/64de3631-c237-4cb6-9b80-aca29c50d694-webhook-cert\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.434044 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.434213 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.434526 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a30d4c31-5122-48fa-8af7-1eba7d2f4254-certs\") pod \"machine-config-server-qbx5c\" (UID: \"a30d4c31-5122-48fa-8af7-1eba7d2f4254\") " pod="openshift-machine-config-operator/machine-config-server-qbx5c" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.438268 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/66a777fc-e931-46d8-bf30-a0ab7f334181-proxy-tls\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.438309 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.438672 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbrcl\" (UniqueName: \"kubernetes.io/projected/e17f16e4-ac2a-417a-a657-edb90df6622c-kube-api-access-sbrcl\") pod \"kube-storage-version-migrator-operator-b67b599dd-p656g\" (UID: \"e17f16e4-ac2a-417a-a657-edb90df6622c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.438882 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0ccc792-7011-435d-934e-ad2d9d94ab90-serving-cert\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.443357 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh4pq\" (UniqueName: \"kubernetes.io/projected/7d2b41be-9ca2-4c6e-a643-23265bdb2ec8-kube-api-access-hh4pq\") pod \"openshift-controller-manager-operator-756b6f6bc6-7bbkt\" (UID: \"7d2b41be-9ca2-4c6e-a643-23265bdb2ec8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.446112 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fxlf\" (UniqueName: \"kubernetes.io/projected/d31847cb-ea3f-46b1-ad2a-12d87b23d882-kube-api-access-6fxlf\") pod \"marketplace-operator-79b997595-wd9j2\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.452870 4823 generic.go:334] "Generic (PLEG): container finished" podID="65497b00-f979-4bad-a290-0af7be716ace" containerID="b330e29c0052954cd941346f25699547e20b0e27012344e4d9e728cce3286ba0" exitCode=0 Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.453002 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" event={"ID":"65497b00-f979-4bad-a290-0af7be716ace","Type":"ContainerDied","Data":"b330e29c0052954cd941346f25699547e20b0e27012344e4d9e728cce3286ba0"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.453041 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" event={"ID":"65497b00-f979-4bad-a290-0af7be716ace","Type":"ContainerStarted","Data":"565c9a6f73e3a79c26a998cc625ced394c6573c3aa778419a5036700745513f2"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.456548 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" event={"ID":"c4256007-3ebf-4991-89b5-83eb7104b8a6","Type":"ContainerStarted","Data":"64303fea76567f80ec24e4f75e1a6541a24437d524eb0579de67424a19d42f98"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.458953 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lndmn" event={"ID":"71d31547-2185-4113-a21e-2f0f2ae4423c","Type":"ContainerStarted","Data":"d7c4ba37356319c90e5c22f8b21f29a43b38c6a7ff498b443c3982927708c7bf"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.465536 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" event={"ID":"dbf97305-8126-4377-801a-3b57bea03ea3","Type":"ContainerStarted","Data":"d616e56e3496b8d47d765da893a008699fbd03d98f81c159dbc43b6c2c43ca0a"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.465576 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" event={"ID":"dbf97305-8126-4377-801a-3b57bea03ea3","Type":"ContainerStarted","Data":"4a87370cfc9c5472b3f97c098f466449455bc896915adcb5f573e2236b3dd504"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.465876 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:50 crc kubenswrapper[4823]: E0320 09:00:50.467513 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:50.967492018 +0000 UTC m=+217.653388392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.478646 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gkgx\" (UniqueName: \"kubernetes.io/projected/a30d4c31-5122-48fa-8af7-1eba7d2f4254-kube-api-access-9gkgx\") pod \"machine-config-server-qbx5c\" (UID: \"a30d4c31-5122-48fa-8af7-1eba7d2f4254\") " pod="openshift-machine-config-operator/machine-config-server-qbx5c" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.479315 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" event={"ID":"8777e543-c661-4a72-9221-e4d398222993","Type":"ContainerStarted","Data":"5cc746746995de849f81c69393cfa3f610ac479e49f0c4be3c72ec3406af7fbb"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.482815 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lz4n9" event={"ID":"e104c875-7420-45b3-b675-a63b92c454e4","Type":"ContainerStarted","Data":"9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.482857 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lz4n9" event={"ID":"e104c875-7420-45b3-b675-a63b92c454e4","Type":"ContainerStarted","Data":"006f746577c69e05f0160fea69f9a67239dcac7a2446d93eac85c8f7fcd34af2"} Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.492685 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhv6m\" (UniqueName: \"kubernetes.io/projected/1b7e18eb-6077-432c-a3b0-383a51b9857d-kube-api-access-jhv6m\") pod \"olm-operator-6b444d44fb-9r5tj\" (UID: \"1b7e18eb-6077-432c-a3b0-383a51b9857d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.533918 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9f8g\" (UniqueName: \"kubernetes.io/projected/1a8b9996-59a1-4167-8943-2c8e66cd1d3c-kube-api-access-w9f8g\") pod \"dns-default-76jgq\" (UID: \"1a8b9996-59a1-4167-8943-2c8e66cd1d3c\") " pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.537708 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qbx5c" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.538252 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.543594 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9pvw\" (UniqueName: \"kubernetes.io/projected/b1d558f6-5989-4d9e-8cfb-d5bd47b42948-kube-api-access-p9pvw\") pod \"multus-admission-controller-857f4d67dd-rfbmh\" (UID: \"b1d558f6-5989-4d9e-8cfb-d5bd47b42948\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.547901 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.555295 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.560546 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl9dg\" (UniqueName: \"kubernetes.io/projected/e0ccc792-7011-435d-934e-ad2d9d94ab90-kube-api-access-jl9dg\") pod \"controller-manager-879f6c89f-p2lfv\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.565093 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.569975 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: E0320 09:00:50.572724 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:51.072711586 +0000 UTC m=+217.758607960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.575123 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.618684 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w85l6\" (UniqueName: \"kubernetes.io/projected/362283f3-bd9a-4c9e-8dbb-b4b3358ed5de-kube-api-access-w85l6\") pod \"service-ca-operator-777779d784-7qskq\" (UID: \"362283f3-bd9a-4c9e-8dbb-b4b3358ed5de\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.633334 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2r2b\" (UniqueName: \"kubernetes.io/projected/66cb5542-0914-40bf-b27d-115a56b07012-kube-api-access-b2r2b\") pod \"oauth-openshift-558db77b4-hlgq9\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.641375 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.652075 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.665168 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qfjh\" (UniqueName: \"kubernetes.io/projected/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-kube-api-access-9qfjh\") pod \"collect-profiles-29566620-smfx8\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.672276 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:50 crc kubenswrapper[4823]: E0320 09:00:50.672781 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:51.172760402 +0000 UTC m=+217.858656786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.675717 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd"] Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.681004 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt2wq\" (UniqueName: \"kubernetes.io/projected/c4aad9d5-b289-45b4-974c-b83bcf1b900f-kube-api-access-xt2wq\") pod \"service-ca-9c57cc56f-zdc26\" (UID: \"c4aad9d5-b289-45b4-974c-b83bcf1b900f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.682049 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvqg5\" (UniqueName: \"kubernetes.io/projected/6e5cca93-e5bb-4297-bf2d-88af3ca984d2-kube-api-access-qvqg5\") pod \"catalog-operator-68c6474976-jjkbd\" (UID: \"6e5cca93-e5bb-4297-bf2d-88af3ca984d2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.693367 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncldg\" (UniqueName: \"kubernetes.io/projected/65a02b02-7f51-4774-b050-292c3fd4f70b-kube-api-access-ncldg\") pod \"auto-csr-approver-29566620-4lfkt\" (UID: \"65a02b02-7f51-4774-b050-292c3fd4f70b\") " pod="openshift-infra/auto-csr-approver-29566620-4lfkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.716391 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57tcj\" (UniqueName: \"kubernetes.io/projected/3e566079-aa6d-4037-a58b-efc6721bc079-kube-api-access-57tcj\") pod \"csi-hostpathplugin-6cfw8\" (UID: \"3e566079-aa6d-4037-a58b-efc6721bc079\") " pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.732767 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt265\" (UniqueName: \"kubernetes.io/projected/6d934348-d496-4ffa-8220-4c7d592a59ce-kube-api-access-dt265\") pod \"ingress-canary-cqddq\" (UID: \"6d934348-d496-4ffa-8220-4c7d592a59ce\") " pod="openshift-ingress-canary/ingress-canary-cqddq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.734997 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qsxgq"] Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.735808 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn"] Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.735831 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.743543 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfvq2\" (UniqueName: \"kubernetes.io/projected/64de3631-c237-4cb6-9b80-aca29c50d694-kube-api-access-rfvq2\") pod \"packageserver-d55dfcdfc-cgkx8\" (UID: \"64de3631-c237-4cb6-9b80-aca29c50d694\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.755941 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.756014 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.756824 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2"] Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.771114 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.772087 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmphk\" (UniqueName: \"kubernetes.io/projected/66a777fc-e931-46d8-bf30-a0ab7f334181-kube-api-access-lmphk\") pod \"machine-config-operator-74547568cd-q7h9k\" (UID: \"66a777fc-e931-46d8-bf30-a0ab7f334181\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.773489 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: E0320 09:00:50.773986 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:51.27396998 +0000 UTC m=+217.959866354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.776126 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k2hn\" (UniqueName: \"kubernetes.io/projected/1d45028a-7db5-4071-901d-2c392d6730b6-kube-api-access-5k2hn\") pod \"package-server-manager-789f6589d5-g2xtb\" (UID: \"1d45028a-7db5-4071-901d-2c392d6730b6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.778958 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.789505 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.801012 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.813643 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.823327 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cqddq" Mar 20 09:00:50 crc kubenswrapper[4823]: W0320 09:00:50.832487 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf49fb4d8_ad57_4153_bd98_6b21c039b971.slice/crio-eb171c6ebcd7f3b99595b5fc2337afd41a7acfd156be4b7ce83ffc4260592f99 WatchSource:0}: Error finding container eb171c6ebcd7f3b99595b5fc2337afd41a7acfd156be4b7ce83ffc4260592f99: Status 404 returned error can't find the container with id eb171c6ebcd7f3b99595b5fc2337afd41a7acfd156be4b7ce83ffc4260592f99 Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.843245 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.874812 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:50 crc kubenswrapper[4823]: E0320 09:00:50.876721 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:51.37669858 +0000 UTC m=+218.062594964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.897040 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.903773 4823 patch_prober.go:28] interesting pod/router-default-5444994796-lndmn container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.903899 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lndmn" podUID="71d31547-2185-4113-a21e-2f0f2ae4423c" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.956437 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.978092 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:50 crc kubenswrapper[4823]: E0320 09:00:50.978465 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:51.478445014 +0000 UTC m=+218.164341388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.978836 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566620-4lfkt" Mar 20 09:00:50 crc kubenswrapper[4823]: I0320 09:00:50.997387 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.009190 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.058052 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.060790 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zdsvr"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.080053 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:51 crc kubenswrapper[4823]: E0320 09:00:51.095806 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:51.595764296 +0000 UTC m=+218.281660670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:51 crc kubenswrapper[4823]: W0320 09:00:51.137160 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d2b41be_9ca2_4c6e_a643_23265bdb2ec8.slice/crio-6aeb06c61d6004c8fc66cf6279bf76adf7be925c093f1f48a3159dd1be1787ff WatchSource:0}: Error finding container 6aeb06c61d6004c8fc66cf6279bf76adf7be925c093f1f48a3159dd1be1787ff: Status 404 returned error can't find the container with id 6aeb06c61d6004c8fc66cf6279bf76adf7be925c093f1f48a3159dd1be1787ff Mar 20 09:00:51 crc kubenswrapper[4823]: W0320 09:00:51.184941 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bdc9b52_2233_463b_ab15_444987a74673.slice/crio-4755f4b2b6172bcd7fe779d8a71836a428dd3db871cb154a76cc4f85f285f8e8 WatchSource:0}: Error finding container 4755f4b2b6172bcd7fe779d8a71836a428dd3db871cb154a76cc4f85f285f8e8: Status 404 returned error can't find the container with id 4755f4b2b6172bcd7fe779d8a71836a428dd3db871cb154a76cc4f85f285f8e8 Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.185778 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:51 crc kubenswrapper[4823]: E0320 09:00:51.186347 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:51.686331021 +0000 UTC m=+218.372227395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.277936 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" podStartSLOduration=175.277914784 podStartE2EDuration="2m55.277914784s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:51.276211697 +0000 UTC m=+217.962108071" watchObservedRunningTime="2026-03-20 09:00:51.277914784 +0000 UTC m=+217.963811158" Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.289453 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:51 crc kubenswrapper[4823]: E0320 09:00:51.289767 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:51.78975109 +0000 UTC m=+218.475647454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.308346 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-grjk6" podStartSLOduration=176.308327412 podStartE2EDuration="2m56.308327412s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:51.306422909 +0000 UTC m=+217.992319363" watchObservedRunningTime="2026-03-20 09:00:51.308327412 +0000 UTC m=+217.994223786" Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.333664 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.343397 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rfbmh"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.350598 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.390532 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:51 crc kubenswrapper[4823]: E0320 09:00:51.390924 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:51.890909947 +0000 UTC m=+218.576806321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.408733 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-lndmn" podStartSLOduration=176.408709927 podStartE2EDuration="2m56.408709927s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:51.394647979 +0000 UTC m=+218.080544353" watchObservedRunningTime="2026-03-20 09:00:51.408709927 +0000 UTC m=+218.094606301" Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.413598 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.416711 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wd9j2"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.491078 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:51 crc kubenswrapper[4823]: E0320 09:00:51.491354 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:51.991333604 +0000 UTC m=+218.677229978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.499378 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lndmn" event={"ID":"71d31547-2185-4113-a21e-2f0f2ae4423c","Type":"ContainerStarted","Data":"2589ca11a58c03be1d91802763678e5a24224c7061539482ad8fb2f70c762853"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.524776 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" event={"ID":"249a7568-39b3-4332-a847-67a1b08d00bc","Type":"ContainerStarted","Data":"b7e4a654c03f3874c251a8ec1d5d04f6a890a73d4c96eeaf9944e3111044cdb6"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.524822 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" event={"ID":"249a7568-39b3-4332-a847-67a1b08d00bc","Type":"ContainerStarted","Data":"056360627aff391b4848b64afbcf90c570f8d70751ab0da5260ed1b0c5687ab9"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.527178 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" event={"ID":"f49fb4d8-ad57-4153-bd98-6b21c039b971","Type":"ContainerStarted","Data":"eb171c6ebcd7f3b99595b5fc2337afd41a7acfd156be4b7ce83ffc4260592f99"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.535202 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qbx5c" event={"ID":"a30d4c31-5122-48fa-8af7-1eba7d2f4254","Type":"ContainerStarted","Data":"737acc14cc8e64acf38062f878335c6b183242161f4290baab95af709df00a20"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.539071 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" event={"ID":"7bdc9b52-2233-463b-ab15-444987a74673","Type":"ContainerStarted","Data":"4755f4b2b6172bcd7fe779d8a71836a428dd3db871cb154a76cc4f85f285f8e8"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.542090 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64" event={"ID":"4da55054-ed9d-4d06-a94c-5f225541f7ea","Type":"ContainerStarted","Data":"8232fc320523e3dc9fe99dbe689c0c39bd3f5ecaf198fd6eb5f806a051d51c13"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.542110 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64" event={"ID":"4da55054-ed9d-4d06-a94c-5f225541f7ea","Type":"ContainerStarted","Data":"68f3a96ee03ff77adb7a45dea40fd3e733a71d6ddebbe13b3fc9b7919e9e30e4"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.549298 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.555244 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" event={"ID":"5b3679cb-16fd-4774-88a5-15d92fe16a43","Type":"ContainerStarted","Data":"eb7f15c37a7f25ea812d8cc914a158c2df3d9343089ac783232caec827db0242"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.561457 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qsxgq" event={"ID":"77917c9c-14bd-4653-ac86-571eda870a4c","Type":"ContainerStarted","Data":"c6d323931ad504b33f135c7ab51438bfdc15fbe7194039be86ac96f30d114931"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.566525 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.567618 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" event={"ID":"2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2","Type":"ContainerStarted","Data":"31d0b8b91f0a83ca37b1bcdf9e3d47748fc54b6f8289bbab452a1f57867b33ab"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.567664 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" event={"ID":"2ec8d6a2-df1b-4d7c-9d3d-9ee73a6605a2","Type":"ContainerStarted","Data":"0abc91cbe7ae3054accefceda1c73d4a8c79146b05dffc4d5da92033d5159009"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.572607 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" event={"ID":"45124dc2-68b5-411c-9826-be34d809a370","Type":"ContainerStarted","Data":"613611ae697f09d7917d284ece93501f85962bb8b17a76ab9dff73c3242163dc"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.573879 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" event={"ID":"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6","Type":"ContainerStarted","Data":"1a43f301f9cf73d97cec927fd231a1ecef8ddef73b5cecb404d0bf1ca0cc8fe7"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.584263 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" event={"ID":"dbf97305-8126-4377-801a-3b57bea03ea3","Type":"ContainerStarted","Data":"1f1afe5d19849d5ba3f64ffb6bc54bb5d84acfdd207eca95c707253a746e9ab2"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.587832 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" event={"ID":"7d2b41be-9ca2-4c6e-a643-23265bdb2ec8","Type":"ContainerStarted","Data":"6aeb06c61d6004c8fc66cf6279bf76adf7be925c093f1f48a3159dd1be1787ff"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.592717 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:51 crc kubenswrapper[4823]: E0320 09:00:51.593068 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.093052366 +0000 UTC m=+218.778948740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.607243 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7qskq"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.613870 4823 generic.go:334] "Generic (PLEG): container finished" podID="70a71b7a-1d65-4ca4-a7c5-25a582208c0b" containerID="bf22dc77260b223912e747f5f23fd29e9976041e4ba270bcb08ee407e9ccc8b8" exitCode=0 Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.613946 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sswrk" event={"ID":"70a71b7a-1d65-4ca4-a7c5-25a582208c0b","Type":"ContainerDied","Data":"bf22dc77260b223912e747f5f23fd29e9976041e4ba270bcb08ee407e9ccc8b8"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.617371 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" event={"ID":"8580a746-c68f-4b57-8f69-f31fb855b785","Type":"ContainerStarted","Data":"7ec2db8a8978182aaea275f269f52400fc3c0dfbcf807f99f2224b6fdef95225"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.617411 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" event={"ID":"8580a746-c68f-4b57-8f69-f31fb855b785","Type":"ContainerStarted","Data":"61f65f13a70ae7e22bb88b2913376e8f392e88e6869ea38c89f09b7d33393b1a"} Mar 20 09:00:51 crc kubenswrapper[4823]: W0320 09:00:51.620571 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda391e636_89df_41fd_9148_9500306e1e0c.slice/crio-ae907a3e8bf89a234008dc3c0e4c28808fe46fbc0113b14b6ec1dee203f7522a WatchSource:0}: Error finding container ae907a3e8bf89a234008dc3c0e4c28808fe46fbc0113b14b6ec1dee203f7522a: Status 404 returned error can't find the container with id ae907a3e8bf89a234008dc3c0e4c28808fe46fbc0113b14b6ec1dee203f7522a Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.650158 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" event={"ID":"65497b00-f979-4bad-a290-0af7be716ace","Type":"ContainerStarted","Data":"21a6e0927303f91c8b7ab02739fbc7e5206f7204cd4ac767d93266d8188cff9b"} Mar 20 09:00:51 crc kubenswrapper[4823]: W0320 09:00:51.666288 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode17f16e4_ac2a_417a_a657_edb90df6622c.slice/crio-cb04e88ecb0a7eeb63829efd41472a2c74f33a72a64ef55665fbb21182b010bf WatchSource:0}: Error finding container cb04e88ecb0a7eeb63829efd41472a2c74f33a72a64ef55665fbb21182b010bf: Status 404 returned error can't find the container with id cb04e88ecb0a7eeb63829efd41472a2c74f33a72a64ef55665fbb21182b010bf Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.671862 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" event={"ID":"c4256007-3ebf-4991-89b5-83eb7104b8a6","Type":"ContainerStarted","Data":"5586f55ba73d6adbd0aeaa2beb1046da7289bdefa4779a0d8e1ba516bd24d900"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.671911 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" event={"ID":"c4256007-3ebf-4991-89b5-83eb7104b8a6","Type":"ContainerStarted","Data":"b150b598e7531b6a1b28c1d675cc4a8ae883b177a0e2b9e8160976a87afc07d2"} Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.680209 4823 patch_prober.go:28] interesting pod/downloads-7954f5f757-grjk6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.680259 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grjk6" podUID="98f13ee0-723b-4ef5-851b-2665009a9d47" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.683163 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-lz4n9" podStartSLOduration=176.683145877 podStartE2EDuration="2m56.683145877s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:51.634688733 +0000 UTC m=+218.320585127" watchObservedRunningTime="2026-03-20 09:00:51.683145877 +0000 UTC m=+218.369042251" Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.687543 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8"] Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.693979 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.695183 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:51 crc kubenswrapper[4823]: E0320 09:00:51.695237 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.19522364 +0000 UTC m=+218.881120004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.696395 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:51 crc kubenswrapper[4823]: W0320 09:00:51.698993 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd31847cb_ea3f_46b1_ad2a_12d87b23d882.slice/crio-3905787e3e265896fa9010989ce5bba4300bda58fb310d3a1c270f7c7eae46f2 WatchSource:0}: Error finding container 3905787e3e265896fa9010989ce5bba4300bda58fb310d3a1c270f7c7eae46f2: Status 404 returned error can't find the container with id 3905787e3e265896fa9010989ce5bba4300bda58fb310d3a1c270f7c7eae46f2 Mar 20 09:00:51 crc kubenswrapper[4823]: E0320 09:00:51.701674 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.201657698 +0000 UTC m=+218.887554072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.799368 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:51 crc kubenswrapper[4823]: E0320 09:00:51.799684 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.299667408 +0000 UTC m=+218.985563782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.900553 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:51 crc kubenswrapper[4823]: E0320 09:00:51.901359 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.401347169 +0000 UTC m=+219.087243543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.909991 4823 patch_prober.go:28] interesting pod/router-default-5444994796-lndmn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:51 crc kubenswrapper[4823]: [-]has-synced failed: reason withheld Mar 20 09:00:51 crc kubenswrapper[4823]: [+]process-running ok Mar 20 09:00:51 crc kubenswrapper[4823]: healthz check failed Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.910050 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lndmn" podUID="71d31547-2185-4113-a21e-2f0f2ae4423c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:51 crc kubenswrapper[4823]: I0320 09:00:51.974900 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-jzzk2" podStartSLOduration=176.974882085 podStartE2EDuration="2m56.974882085s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:51.974323699 +0000 UTC m=+218.660220073" watchObservedRunningTime="2026-03-20 09:00:51.974882085 +0000 UTC m=+218.660778459" Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.010424 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.010683 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.51066887 +0000 UTC m=+219.196565244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.027780 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k"] Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.040298 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cqddq"] Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.041350 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p2lfv"] Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.114608 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qm6tg" podStartSLOduration=177.114588643 podStartE2EDuration="2m57.114588643s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:52.111507058 +0000 UTC m=+218.797403432" watchObservedRunningTime="2026-03-20 09:00:52.114588643 +0000 UTC m=+218.800485017" Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.116080 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.116507 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.616492917 +0000 UTC m=+219.302389291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.152665 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hlgq9"] Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.217101 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.217279 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.717250992 +0000 UTC m=+219.403147366 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.217387 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.217670 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.717659243 +0000 UTC m=+219.403555617 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.320995 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.321400 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.82136364 +0000 UTC m=+219.507260014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.321639 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.321979 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.821964337 +0000 UTC m=+219.507860711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.343469 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb"] Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.350958 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-76jgq"] Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.369569 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd"] Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.425248 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.425551 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:52.92553549 +0000 UTC m=+219.611431864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.427156 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-mkqpg" podStartSLOduration=177.427140395 podStartE2EDuration="2m57.427140395s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:52.426311952 +0000 UTC m=+219.112208346" watchObservedRunningTime="2026-03-20 09:00:52.427140395 +0000 UTC m=+219.113036769" Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.526686 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.527055 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:53.027042057 +0000 UTC m=+219.712938431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.608595 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zdc26"] Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.621154 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6cfw8"] Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.627881 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.628023 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:53.128000258 +0000 UTC m=+219.813896632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.628061 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.628375 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:53.128367838 +0000 UTC m=+219.814264212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.636346 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-49lkt" podStartSLOduration=176.636326747 podStartE2EDuration="2m56.636326747s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:52.633365576 +0000 UTC m=+219.319261950" watchObservedRunningTime="2026-03-20 09:00:52.636326747 +0000 UTC m=+219.322223121" Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.667240 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vn8gj" podStartSLOduration=177.667221938 podStartE2EDuration="2m57.667221938s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:52.666540699 +0000 UTC m=+219.352437083" watchObservedRunningTime="2026-03-20 09:00:52.667221938 +0000 UTC m=+219.353118312" Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.677248 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" event={"ID":"e17f16e4-ac2a-417a-a657-edb90df6622c","Type":"ContainerStarted","Data":"cb04e88ecb0a7eeb63829efd41472a2c74f33a72a64ef55665fbb21182b010bf"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.678447 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" event={"ID":"7d2b41be-9ca2-4c6e-a643-23265bdb2ec8","Type":"ContainerStarted","Data":"c9f1db3060a88dd4110752bedc75d70b7a1729a01f276e8aca4a99baf0988ce1"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.679626 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" event={"ID":"362283f3-bd9a-4c9e-8dbb-b4b3358ed5de","Type":"ContainerStarted","Data":"47865d7016e2c55b513f45756903ecd5522da6c1ecc84a3e13b045293b749c33"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.681408 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" event={"ID":"f49fb4d8-ad57-4153-bd98-6b21c039b971","Type":"ContainerStarted","Data":"fa40327c1347592fc77c93ddd3573ec51a1db5609d7ea0d168f0ab5c2b529030"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.682191 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" event={"ID":"66cb5542-0914-40bf-b27d-115a56b07012","Type":"ContainerStarted","Data":"80ab244d539acc8e64fddf414d68a41c53f520ad8ef72603dfed045bc3de3088"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.683137 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cqddq" event={"ID":"6d934348-d496-4ffa-8220-4c7d592a59ce","Type":"ContainerStarted","Data":"54924d10b60bf364de329adf493f91662a4ffe25a64001129801ee64428c977a"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.683962 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" event={"ID":"a391e636-89df-41fd-9148-9500306e1e0c","Type":"ContainerStarted","Data":"ae907a3e8bf89a234008dc3c0e4c28808fe46fbc0113b14b6ec1dee203f7522a"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.684942 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" event={"ID":"66a777fc-e931-46d8-bf30-a0ab7f334181","Type":"ContainerStarted","Data":"bc947489d5037b87750be1ff470aa442d397a70e19aa1dbd790c67751dbdf9c5"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.685794 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" event={"ID":"1b7e18eb-6077-432c-a3b0-383a51b9857d","Type":"ContainerStarted","Data":"f0c7e31e353ae1da1cdec9744dfec8b54bc75d5bac7eaf4f33d7810fe51250e0"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.686589 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" event={"ID":"64de3631-c237-4cb6-9b80-aca29c50d694","Type":"ContainerStarted","Data":"89685180df514576dfc2939250708305dd45e129f9011e451ac024afd42568db"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.687872 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qbx5c" event={"ID":"a30d4c31-5122-48fa-8af7-1eba7d2f4254","Type":"ContainerStarted","Data":"ee408d7708d33429f786cd56470d15f79f681931e4fd7909c8513707c27b27f2"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.688643 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" event={"ID":"d31847cb-ea3f-46b1-ad2a-12d87b23d882","Type":"ContainerStarted","Data":"3905787e3e265896fa9010989ce5bba4300bda58fb310d3a1c270f7c7eae46f2"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.689959 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qsxgq" event={"ID":"77917c9c-14bd-4653-ac86-571eda870a4c","Type":"ContainerStarted","Data":"4744cb39559557307080b36e2bffbf7d689be626a7ce736b53922f3ce8009bb2"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.690777 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" event={"ID":"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7","Type":"ContainerStarted","Data":"a7b50c5fd9f539a08f3a442445fbae533a4aecc59fbcc67adeeef7988a164e8d"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.691802 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" event={"ID":"45124dc2-68b5-411c-9826-be34d809a370","Type":"ContainerStarted","Data":"f01a42b5e879153ceed91126590d775a82d196533300507eae9c95516bef9c13"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.692510 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x" event={"ID":"b54b2f8f-1e5a-44cb-ac82-77040375eb02","Type":"ContainerStarted","Data":"fd81cd414e2002b812af5edb1a2ba5ffd7fc03af6e86638acf1a29e109348fca"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.693347 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" event={"ID":"b1d558f6-5989-4d9e-8cfb-d5bd47b42948","Type":"ContainerStarted","Data":"ce67d0f5968ee55db2192d84b31cd2b13c17b6370081b3e6633702901c85a410"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.694391 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" event={"ID":"e0ccc792-7011-435d-934e-ad2d9d94ab90","Type":"ContainerStarted","Data":"dcc68970de824bc41ee6f33879a2a204dd30f70134a32531d3d6f788539b9bbf"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.695778 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" event={"ID":"e64e0f18-5f66-49ca-9cca-b97bde4f2dc6","Type":"ContainerStarted","Data":"9a8fe12032031919ccf22929a66f2c85ac0670d6353bbe95d098db4b965504c9"} Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.705597 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7b46d" podStartSLOduration=177.705578855 podStartE2EDuration="2m57.705578855s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:52.704113775 +0000 UTC m=+219.390010149" watchObservedRunningTime="2026-03-20 09:00:52.705578855 +0000 UTC m=+219.391475249" Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.729586 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.729937 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:53.229923185 +0000 UTC m=+219.915819559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.795896 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tbm9g" podStartSLOduration=176.795860352 podStartE2EDuration="2m56.795860352s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:52.747979283 +0000 UTC m=+219.433875667" watchObservedRunningTime="2026-03-20 09:00:52.795860352 +0000 UTC m=+219.481756726" Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.831758 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.834512 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:53.334485756 +0000 UTC m=+220.020382340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.835811 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" podStartSLOduration=176.835782482 podStartE2EDuration="2m56.835782482s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:52.835119984 +0000 UTC m=+219.521016358" watchObservedRunningTime="2026-03-20 09:00:52.835782482 +0000 UTC m=+219.521678856" Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.875637 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qst64" podStartSLOduration=176.875609099 podStartE2EDuration="2m56.875609099s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:52.872090133 +0000 UTC m=+219.557986527" watchObservedRunningTime="2026-03-20 09:00:52.875609099 +0000 UTC m=+219.561505483" Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.900662 4823 patch_prober.go:28] interesting pod/router-default-5444994796-lndmn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:52 crc kubenswrapper[4823]: [-]has-synced failed: reason withheld Mar 20 09:00:52 crc kubenswrapper[4823]: [+]process-running ok Mar 20 09:00:52 crc kubenswrapper[4823]: healthz check failed Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.900736 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lndmn" podUID="71d31547-2185-4113-a21e-2f0f2ae4423c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.910277 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xw7gs" podStartSLOduration=177.908715042 podStartE2EDuration="2m57.908715042s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:52.905494503 +0000 UTC m=+219.591390877" watchObservedRunningTime="2026-03-20 09:00:52.908715042 +0000 UTC m=+219.594611416" Mar 20 09:00:52 crc kubenswrapper[4823]: I0320 09:00:52.933569 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:52 crc kubenswrapper[4823]: E0320 09:00:52.934067 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:53.43404458 +0000 UTC m=+220.119940954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.035568 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.036121 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:53.536097 +0000 UTC m=+220.221993374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.137024 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.137600 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:53.637576616 +0000 UTC m=+220.323473000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.256941 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:53.756714408 +0000 UTC m=+220.442610822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.257541 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.285853 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566620-4lfkt"] Mar 20 09:00:53 crc kubenswrapper[4823]: W0320 09:00:53.329157 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a8b9996_59a1_4167_8943_2c8e66cd1d3c.slice/crio-7c1890823bb61e2cc00bc5758445bbe1d63ee2980aee00e049492ea66ff20689 WatchSource:0}: Error finding container 7c1890823bb61e2cc00bc5758445bbe1d63ee2980aee00e049492ea66ff20689: Status 404 returned error can't find the container with id 7c1890823bb61e2cc00bc5758445bbe1d63ee2980aee00e049492ea66ff20689 Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.358788 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.359177 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:53.859154721 +0000 UTC m=+220.545051115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: W0320 09:00:53.362238 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65a02b02_7f51_4774_b050_292c3fd4f70b.slice/crio-0dc23f2f6bb8904a31284e767413949b1f157d46ea17e175f7e2e9db9585cc69 WatchSource:0}: Error finding container 0dc23f2f6bb8904a31284e767413949b1f157d46ea17e175f7e2e9db9585cc69: Status 404 returned error can't find the container with id 0dc23f2f6bb8904a31284e767413949b1f157d46ea17e175f7e2e9db9585cc69 Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.373150 4823 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.460611 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.461163 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:53.96114092 +0000 UTC m=+220.647037294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.563258 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.563389 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.063361537 +0000 UTC m=+220.749257911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.563577 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.563853 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.06384544 +0000 UTC m=+220.749741814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.665330 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.665802 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.165726246 +0000 UTC m=+220.851622620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.668131 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.668489 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.168471732 +0000 UTC m=+220.854368106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.728011 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" event={"ID":"8580a746-c68f-4b57-8f69-f31fb855b785","Type":"ContainerStarted","Data":"b9e352190847ddb1ce67f1ea3d29307f810d54672802459df3a632ec439484a2"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.756600 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566620-4lfkt" event={"ID":"65a02b02-7f51-4774-b050-292c3fd4f70b","Type":"ContainerStarted","Data":"0dc23f2f6bb8904a31284e767413949b1f157d46ea17e175f7e2e9db9585cc69"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.760835 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" event={"ID":"362283f3-bd9a-4c9e-8dbb-b4b3358ed5de","Type":"ContainerStarted","Data":"3f059c1c644c72c7493c99b07558139bf86af559c65b90845ace252c911c6764"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.765620 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" event={"ID":"1d45028a-7db5-4071-901d-2c392d6730b6","Type":"ContainerStarted","Data":"dcf80a96b5d1aec0367626fa69f5b9c64e0102c44426375e48a53d81c9adec73"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.769375 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.770928 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.270905904 +0000 UTC m=+220.956802278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.778135 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gkv7j" podStartSLOduration=178.778092582 podStartE2EDuration="2m58.778092582s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:53.748416305 +0000 UTC m=+220.434312679" watchObservedRunningTime="2026-03-20 09:00:53.778092582 +0000 UTC m=+220.463988956" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.785470 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" event={"ID":"b1d558f6-5989-4d9e-8cfb-d5bd47b42948","Type":"ContainerStarted","Data":"2211eee14100ae90228055620cb451ca7ac34552ba1cf6df6e8e303c1896951c"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.793925 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" event={"ID":"7bdc9b52-2233-463b-ab15-444987a74673","Type":"ContainerStarted","Data":"7cbf646e3146f507de1268e66260e48a9a24a4247aeaf4ea6c1b8b7e7652b7af"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.802270 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" event={"ID":"e17f16e4-ac2a-417a-a657-edb90df6622c","Type":"ContainerStarted","Data":"d192f7da40a3370cdbe28f98547d444e01e8d9b914d0374757c21745c9d7d3d4"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.808065 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" event={"ID":"6e5cca93-e5bb-4297-bf2d-88af3ca984d2","Type":"ContainerStarted","Data":"2ed4b19fdceaa49a4f9ff464884ce7beabb721ff624c79f042c558ad4f0d5006"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.817250 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7qskq" podStartSLOduration=177.81723072 podStartE2EDuration="2m57.81723072s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:53.781930458 +0000 UTC m=+220.467826822" watchObservedRunningTime="2026-03-20 09:00:53.81723072 +0000 UTC m=+220.503127094" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.831280 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-zdsvr" podStartSLOduration=178.831251986 podStartE2EDuration="2m58.831251986s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:53.823611766 +0000 UTC m=+220.509508140" watchObservedRunningTime="2026-03-20 09:00:53.831251986 +0000 UTC m=+220.517148360" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.848293 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p656g" podStartSLOduration=177.848262045 podStartE2EDuration="2m57.848262045s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:53.843807413 +0000 UTC m=+220.529703787" watchObservedRunningTime="2026-03-20 09:00:53.848262045 +0000 UTC m=+220.534158419" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.873396 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" event={"ID":"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7","Type":"ContainerStarted","Data":"8262abc138eac431152da2dcf732d57c8026e0ca34981101a5a1495833ef3aff"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.873936 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.875220 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.375205467 +0000 UTC m=+221.061101951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.892896 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" event={"ID":"5b3679cb-16fd-4774-88a5-15d92fe16a43","Type":"ContainerStarted","Data":"8dd7d8ed11417071721a5cb93c052609a4edd5af343f0d2ebc51376f65a0776d"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.904911 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" podStartSLOduration=53.904887705 podStartE2EDuration="53.904887705s" podCreationTimestamp="2026-03-20 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:53.901888653 +0000 UTC m=+220.587785027" watchObservedRunningTime="2026-03-20 09:00:53.904887705 +0000 UTC m=+220.590784079" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.910528 4823 patch_prober.go:28] interesting pod/router-default-5444994796-lndmn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:53 crc kubenswrapper[4823]: [-]has-synced failed: reason withheld Mar 20 09:00:53 crc kubenswrapper[4823]: [+]process-running ok Mar 20 09:00:53 crc kubenswrapper[4823]: healthz check failed Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.910601 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lndmn" podUID="71d31547-2185-4113-a21e-2f0f2ae4423c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.917506 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" event={"ID":"3e566079-aa6d-4037-a58b-efc6721bc079","Type":"ContainerStarted","Data":"0b1c8cd604c1fe978b4eca43ed4352ba8346431ca4fbe6baf4bb2a1ffb9d19f3"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.919013 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-76jgq" event={"ID":"1a8b9996-59a1-4167-8943-2c8e66cd1d3c","Type":"ContainerStarted","Data":"7c1890823bb61e2cc00bc5758445bbe1d63ee2980aee00e049492ea66ff20689"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.922807 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" event={"ID":"c4aad9d5-b289-45b4-974c-b83bcf1b900f","Type":"ContainerStarted","Data":"f862fe6fc643aba4fdfc41f9e171ade0c1bf86ed82c05a1e796d9bec9def5e09"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.928472 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" podStartSLOduration=177.928454224 podStartE2EDuration="2m57.928454224s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:53.925667167 +0000 UTC m=+220.611563531" watchObservedRunningTime="2026-03-20 09:00:53.928454224 +0000 UTC m=+220.614350598" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.942662 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" event={"ID":"64de3631-c237-4cb6-9b80-aca29c50d694","Type":"ContainerStarted","Data":"1d5e3e76799975c03b1c8b0dd391251061b24c3b82dca666862ace5658e0cdc3"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.943268 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.946030 4823 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-cgkx8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.946082 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" podUID="64de3631-c237-4cb6-9b80-aca29c50d694" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.952596 4823 generic.go:334] "Generic (PLEG): container finished" podID="f49fb4d8-ad57-4153-bd98-6b21c039b971" containerID="fa40327c1347592fc77c93ddd3573ec51a1db5609d7ea0d168f0ab5c2b529030" exitCode=0 Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.952765 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" event={"ID":"f49fb4d8-ad57-4153-bd98-6b21c039b971","Type":"ContainerDied","Data":"fa40327c1347592fc77c93ddd3573ec51a1db5609d7ea0d168f0ab5c2b529030"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.973514 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" event={"ID":"d31847cb-ea3f-46b1-ad2a-12d87b23d882","Type":"ContainerStarted","Data":"7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67"} Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.974351 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.975175 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" podStartSLOduration=177.975151221 podStartE2EDuration="2m57.975151221s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:53.971424988 +0000 UTC m=+220.657321362" watchObservedRunningTime="2026-03-20 09:00:53.975151221 +0000 UTC m=+220.661047595" Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.975959 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:53 crc kubenswrapper[4823]: E0320 09:00:53.976412 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.476395205 +0000 UTC m=+221.162291579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.977302 4823 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wd9j2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Mar 20 09:00:53 crc kubenswrapper[4823]: I0320 09:00:53.977332 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" podUID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.009399 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sswrk" event={"ID":"70a71b7a-1d65-4ca4-a7c5-25a582208c0b","Type":"ContainerStarted","Data":"ffa9fb49c611e64b22e0e4d76b4c0f93d295f8f7fa680b7eb1b3f36c7de8e857"} Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.014145 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" event={"ID":"1b7e18eb-6077-432c-a3b0-383a51b9857d","Type":"ContainerStarted","Data":"74e2adfee1c73ae38c2d04649d1b5a152dbb705a40fc87e1ccf4b238f293ab6a"} Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.014476 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.038691 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" podStartSLOduration=178.038674211 podStartE2EDuration="2m58.038674211s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:54.018974278 +0000 UTC m=+220.704870652" watchObservedRunningTime="2026-03-20 09:00:54.038674211 +0000 UTC m=+220.724570595" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.039166 4823 patch_prober.go:28] interesting pod/console-operator-58897d9998-qsxgq container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/readyz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.039230 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-qsxgq" podUID="77917c9c-14bd-4653-ac86-571eda870a4c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/readyz\": dial tcp 10.217.0.29:8443: connect: connection refused" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.045443 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-jbsnd" podStartSLOduration=179.045394706 podStartE2EDuration="2m59.045394706s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:54.036997685 +0000 UTC m=+220.722894069" watchObservedRunningTime="2026-03-20 09:00:54.045394706 +0000 UTC m=+220.731291110" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.081602 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.089208 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.589194353 +0000 UTC m=+221.275090727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.113456 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" podStartSLOduration=178.11343115 podStartE2EDuration="2m58.11343115s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:54.064327168 +0000 UTC m=+220.750223552" watchObservedRunningTime="2026-03-20 09:00:54.11343115 +0000 UTC m=+220.799327524" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.138380 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-qsxgq" podStartSLOduration=179.138342976 podStartE2EDuration="2m59.138342976s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:54.13556166 +0000 UTC m=+220.821458034" watchObservedRunningTime="2026-03-20 09:00:54.138342976 +0000 UTC m=+220.824239350" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.143893 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7bbkt" podStartSLOduration=179.143866558 podStartE2EDuration="2m59.143866558s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:54.114390387 +0000 UTC m=+220.800286761" watchObservedRunningTime="2026-03-20 09:00:54.143866558 +0000 UTC m=+220.829762932" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.183837 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-qbx5c" podStartSLOduration=7.183814589 podStartE2EDuration="7.183814589s" podCreationTimestamp="2026-03-20 09:00:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:54.159347695 +0000 UTC m=+220.845244069" watchObservedRunningTime="2026-03-20 09:00:54.183814589 +0000 UTC m=+220.869710963" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.186695 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.187130 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.68711069 +0000 UTC m=+221.373007064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.225295 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pcdfn" podStartSLOduration=179.225274481 podStartE2EDuration="2m59.225274481s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:54.1863822 +0000 UTC m=+220.872278574" watchObservedRunningTime="2026-03-20 09:00:54.225274481 +0000 UTC m=+220.911170865" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.288543 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.288958 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.788944036 +0000 UTC m=+221.474840420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.350652 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.351367 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.357792 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.389829 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.390048 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.890000329 +0000 UTC m=+221.575896713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.390199 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.391427 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.891414008 +0000 UTC m=+221.577310382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.491508 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.491736 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.991704231 +0000 UTC m=+221.677600605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.491945 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.492293 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:54.992285347 +0000 UTC m=+221.678181721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.593149 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.593343 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.093308211 +0000 UTC m=+221.779204595 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.593401 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.593906 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.093895577 +0000 UTC m=+221.779791951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.694322 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.694597 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.19456657 +0000 UTC m=+221.880462944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.694935 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.695315 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.19530473 +0000 UTC m=+221.881201104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.795752 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.795968 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.295937372 +0000 UTC m=+221.981833756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.796062 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.796368 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.296355904 +0000 UTC m=+221.982252288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.838755 4823 ???:1] "http: TLS handshake error from 192.168.126.11:44068: no serving certificate available for the kubelet" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.864700 4823 ???:1] "http: TLS handshake error from 192.168.126.11:44080: no serving certificate available for the kubelet" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.887868 4823 ???:1] "http: TLS handshake error from 192.168.126.11:44096: no serving certificate available for the kubelet" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.897517 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.897781 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.397751157 +0000 UTC m=+222.083647531 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.897893 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.898179 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.398172349 +0000 UTC m=+222.084068723 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.899864 4823 patch_prober.go:28] interesting pod/router-default-5444994796-lndmn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:54 crc kubenswrapper[4823]: [-]has-synced failed: reason withheld Mar 20 09:00:54 crc kubenswrapper[4823]: [+]process-running ok Mar 20 09:00:54 crc kubenswrapper[4823]: healthz check failed Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.899927 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lndmn" podUID="71d31547-2185-4113-a21e-2f0f2ae4423c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.935904 4823 ???:1] "http: TLS handshake error from 192.168.126.11:44100: no serving certificate available for the kubelet" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.961060 4823 ???:1] "http: TLS handshake error from 192.168.126.11:44102: no serving certificate available for the kubelet" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.997662 4823 ???:1] "http: TLS handshake error from 192.168.126.11:44106: no serving certificate available for the kubelet" Mar 20 09:00:54 crc kubenswrapper[4823]: I0320 09:00:54.998923 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:54 crc kubenswrapper[4823]: E0320 09:00:54.999226 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.499201453 +0000 UTC m=+222.185097817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.022281 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" event={"ID":"a391e636-89df-41fd-9148-9500306e1e0c","Type":"ContainerStarted","Data":"ec77f17783f8b0d9e604d580d17db7e665db3588f5d49b20461a430fc6203f7e"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.036140 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" event={"ID":"f49fb4d8-ad57-4153-bd98-6b21c039b971","Type":"ContainerStarted","Data":"0248d2ae7e11ba7df303c426dce8cfc7d16528b6acb91f2439399a7f63c30be8"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.036724 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.039793 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-r4kcw" podStartSLOduration=179.03977475 podStartE2EDuration="2m59.03977475s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.038095844 +0000 UTC m=+221.723992218" watchObservedRunningTime="2026-03-20 09:00:55.03977475 +0000 UTC m=+221.725671124" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.048152 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" event={"ID":"b1d558f6-5989-4d9e-8cfb-d5bd47b42948","Type":"ContainerStarted","Data":"a53cb760f05a22a9d176ce2a5ebce8e21462ebcccfff64f07c97bfd9312fc6b1"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.055669 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" event={"ID":"6e5cca93-e5bb-4297-bf2d-88af3ca984d2","Type":"ContainerStarted","Data":"a5a631c0902532ea958cfeb447a7e11297556b7743e82d821d5e6386e247f0ac"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.056826 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.063620 4823 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-jjkbd container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.063699 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" podUID="6e5cca93-e5bb-4297-bf2d-88af3ca984d2" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.066946 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" event={"ID":"e0ccc792-7011-435d-934e-ad2d9d94ab90","Type":"ContainerStarted","Data":"9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.067009 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" podStartSLOduration=180.06698633 podStartE2EDuration="3m0.06698633s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.057119698 +0000 UTC m=+221.743016072" watchObservedRunningTime="2026-03-20 09:00:55.06698633 +0000 UTC m=+221.752882704" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.067479 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.069536 4823 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-p2lfv container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.069606 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" podUID="e0ccc792-7011-435d-934e-ad2d9d94ab90" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.069642 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cqddq" event={"ID":"6d934348-d496-4ffa-8220-4c7d592a59ce","Type":"ContainerStarted","Data":"a42587fff45d1a68684630fe6d88de6ab5039cf9085f916c103968a047181d6e"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.093396 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-rfbmh" podStartSLOduration=179.093375387 podStartE2EDuration="2m59.093375387s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.086361324 +0000 UTC m=+221.772257698" watchObservedRunningTime="2026-03-20 09:00:55.093375387 +0000 UTC m=+221.779271771" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.097383 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sx7sz" event={"ID":"5b3679cb-16fd-4774-88a5-15d92fe16a43","Type":"ContainerStarted","Data":"f645024bc23dc0542485d19a55cd64b34a18b32d5125bcc0dbdac6380ad52768"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.100815 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" event={"ID":"1d45028a-7db5-4071-901d-2c392d6730b6","Type":"ContainerStarted","Data":"330ff20f487134145c3310e9ed47d9f2b16e25d76325c9cc05dca2026b733497"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.100845 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" event={"ID":"1d45028a-7db5-4071-901d-2c392d6730b6","Type":"ContainerStarted","Data":"c31a4adcfaab766d05dc4a6c8a4e1b5e77f91d876d43551f4ea31286cde10e4b"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.101042 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.101491 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.103178 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.603163387 +0000 UTC m=+222.289059761 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.103863 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x" event={"ID":"b54b2f8f-1e5a-44cb-ac82-77040375eb02","Type":"ContainerStarted","Data":"d42ead28c81f5a0b5133c23c9b9706ff71e72066586f43872d947c02da14b896"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.104899 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x" event={"ID":"b54b2f8f-1e5a-44cb-ac82-77040375eb02","Type":"ContainerStarted","Data":"78e09989b6be60d9435fbafdaa763dc1706ffe142398e6b6ea1afa3eb68ca0a3"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.112873 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" podStartSLOduration=179.112853233 podStartE2EDuration="2m59.112853233s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.110163449 +0000 UTC m=+221.796059813" watchObservedRunningTime="2026-03-20 09:00:55.112853233 +0000 UTC m=+221.798749607" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.132377 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" event={"ID":"66a777fc-e931-46d8-bf30-a0ab7f334181","Type":"ContainerStarted","Data":"8d9a598d1af363879dea8ee220b6734304809f218f917d0eb6f02bf7689738ff"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.132426 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" event={"ID":"66a777fc-e931-46d8-bf30-a0ab7f334181","Type":"ContainerStarted","Data":"070c12f016795d3ee518d2cdf32dd8daac0858f5903bbc48df8231952691614a"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.138084 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-cqddq" podStartSLOduration=8.138064928 podStartE2EDuration="8.138064928s" podCreationTimestamp="2026-03-20 09:00:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.136244048 +0000 UTC m=+221.822140412" watchObservedRunningTime="2026-03-20 09:00:55.138064928 +0000 UTC m=+221.823961302" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.147687 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sswrk" event={"ID":"70a71b7a-1d65-4ca4-a7c5-25a582208c0b","Type":"ContainerStarted","Data":"48c452aea29bcbe406ceef7ac76313ceb03fbaf68138b87bcd58b3922453ccd0"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.161980 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t8r2x" podStartSLOduration=179.161958997 podStartE2EDuration="2m59.161958997s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.159297613 +0000 UTC m=+221.845193987" watchObservedRunningTime="2026-03-20 09:00:55.161958997 +0000 UTC m=+221.847855381" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.167253 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-76jgq" event={"ID":"1a8b9996-59a1-4167-8943-2c8e66cd1d3c","Type":"ContainerStarted","Data":"9f77d3462eb247ac2b9602df4e0c3877a3d337165d0da0ca9db329416400abbf"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.167565 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-76jgq" event={"ID":"1a8b9996-59a1-4167-8943-2c8e66cd1d3c","Type":"ContainerStarted","Data":"de6be92dc7786580d790afefaae070f08065a5cd3d1ba5258858b37f03074876"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.168427 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-76jgq" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.174486 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" event={"ID":"c4aad9d5-b289-45b4-974c-b83bcf1b900f","Type":"ContainerStarted","Data":"a3fa1d7834e28496fc458ecd59926b1e7651c542ecba0744a069681af102e5d1"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.192660 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" event={"ID":"66cb5542-0914-40bf-b27d-115a56b07012","Type":"ContainerStarted","Data":"b521b5eb4d29072d233223caf224b844651693501345b5c86f510b27a5cd6a8d"} Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.193055 4823 patch_prober.go:28] interesting pod/console-operator-58897d9998-qsxgq container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/readyz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.193112 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-qsxgq" podUID="77917c9c-14bd-4653-ac86-571eda870a4c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/readyz\": dial tcp 10.217.0.29:8443: connect: connection refused" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.194114 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.194143 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.194385 4823 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wd9j2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.194437 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" podUID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.195061 4823 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-cgkx8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.195086 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" podUID="64de3631-c237-4cb6-9b80-aca29c50d694" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.198058 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" podStartSLOduration=180.19803152 podStartE2EDuration="3m0.19803152s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.189171856 +0000 UTC m=+221.875068230" watchObservedRunningTime="2026-03-20 09:00:55.19803152 +0000 UTC m=+221.883927894" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.198878 4823 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-hlgq9 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.38:6443/healthz\": dial tcp 10.217.0.38:6443: connect: connection refused" start-of-body= Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.198919 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" podUID="66cb5542-0914-40bf-b27d-115a56b07012" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.38:6443/healthz\": dial tcp 10.217.0.38:6443: connect: connection refused" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.198982 4823 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-9r5tj container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.198994 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" podUID="1b7e18eb-6077-432c-a3b0-383a51b9857d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.206826 4823 ???:1] "http: TLS handshake error from 192.168.126.11:44114: no serving certificate available for the kubelet" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.207534 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r28hg" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.210194 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.210531 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.710498933 +0000 UTC m=+222.396395327 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.211098 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.212602 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" podStartSLOduration=179.212591702 podStartE2EDuration="2m59.212591702s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.2111193 +0000 UTC m=+221.897015684" watchObservedRunningTime="2026-03-20 09:00:55.212591702 +0000 UTC m=+221.898488076" Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.213040 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.713016182 +0000 UTC m=+222.398912626 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.236930 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-76jgq" podStartSLOduration=8.236910201 podStartE2EDuration="8.236910201s" podCreationTimestamp="2026-03-20 09:00:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.234985269 +0000 UTC m=+221.920881643" watchObservedRunningTime="2026-03-20 09:00:55.236910201 +0000 UTC m=+221.922806575" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.258311 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-sswrk" podStartSLOduration=180.258296 podStartE2EDuration="3m0.258296s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.257786806 +0000 UTC m=+221.943683180" watchObservedRunningTime="2026-03-20 09:00:55.258296 +0000 UTC m=+221.944192374" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.278372 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-q7h9k" podStartSLOduration=179.278354853 podStartE2EDuration="2m59.278354853s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.275574236 +0000 UTC m=+221.961470610" watchObservedRunningTime="2026-03-20 09:00:55.278354853 +0000 UTC m=+221.964251227" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.301632 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zdc26" podStartSLOduration=179.301574523 podStartE2EDuration="2m59.301574523s" podCreationTimestamp="2026-03-20 08:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.29786163 +0000 UTC m=+221.983758044" watchObservedRunningTime="2026-03-20 09:00:55.301574523 +0000 UTC m=+221.987470897" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.313481 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.314551 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.814533849 +0000 UTC m=+222.500430223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.372918 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" podStartSLOduration=180.372900627 podStartE2EDuration="3m0.372900627s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:55.3715327 +0000 UTC m=+222.057429074" watchObservedRunningTime="2026-03-20 09:00:55.372900627 +0000 UTC m=+222.058797001" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.415634 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.415994 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:55.915982184 +0000 UTC m=+222.601878558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.516340 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.516578 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.016548535 +0000 UTC m=+222.702444909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.516664 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.516962 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.016950246 +0000 UTC m=+222.702846620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.617141 4823 ???:1] "http: TLS handshake error from 192.168.126.11:44130: no serving certificate available for the kubelet" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.617732 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.617931 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.117909418 +0000 UTC m=+222.803805802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.618047 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.618327 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.118318269 +0000 UTC m=+222.804214643 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.719953 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.720178 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.220153094 +0000 UTC m=+222.906049458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.721177 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.721667 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.221650816 +0000 UTC m=+222.907547410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.822787 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.823007 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.322972927 +0000 UTC m=+223.008869291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.823060 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.823522 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.323506261 +0000 UTC m=+223.009402635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.904609 4823 patch_prober.go:28] interesting pod/router-default-5444994796-lndmn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:55 crc kubenswrapper[4823]: [-]has-synced failed: reason withheld Mar 20 09:00:55 crc kubenswrapper[4823]: [+]process-running ok Mar 20 09:00:55 crc kubenswrapper[4823]: healthz check failed Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.905054 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lndmn" podUID="71d31547-2185-4113-a21e-2f0f2ae4423c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:55 crc kubenswrapper[4823]: I0320 09:00:55.924180 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:55 crc kubenswrapper[4823]: E0320 09:00:55.924516 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.424497573 +0000 UTC m=+223.110393947 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.025485 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.025862 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.525846986 +0000 UTC m=+223.211743360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.126657 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.127057 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.627041683 +0000 UTC m=+223.312938057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.225039 4823 generic.go:334] "Generic (PLEG): container finished" podID="4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7" containerID="8262abc138eac431152da2dcf732d57c8026e0ca34981101a5a1495833ef3aff" exitCode=0 Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.226155 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" event={"ID":"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7","Type":"ContainerDied","Data":"8262abc138eac431152da2dcf732d57c8026e0ca34981101a5a1495833ef3aff"} Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.227693 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.228019 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.728007595 +0000 UTC m=+223.413903969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.250973 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" event={"ID":"3e566079-aa6d-4037-a58b-efc6721bc079","Type":"ContainerStarted","Data":"7b649441c24bde1003573536c1e72b0009a9d352dc637d428350049a1049b54a"} Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.252879 4823 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-p2lfv container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.252916 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" podUID="e0ccc792-7011-435d-934e-ad2d9d94ab90" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.258122 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.269313 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9r5tj" Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.277346 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jjkbd" Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.314838 4823 ???:1] "http: TLS handshake error from 192.168.126.11:44136: no serving certificate available for the kubelet" Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.328519 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.330083 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.830064266 +0000 UTC m=+223.515960630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.431394 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.431721 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:56.931709877 +0000 UTC m=+223.617606251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.533081 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.533473 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.03345417 +0000 UTC m=+223.719350544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.634149 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.634530 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.134514724 +0000 UTC m=+223.820411098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.735629 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.735867 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.235840226 +0000 UTC m=+223.921736600 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.735996 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.736408 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.236395851 +0000 UTC m=+223.922292225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.759687 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.837141 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.837388 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.337352542 +0000 UTC m=+224.023248916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.843613 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.843974 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.343958084 +0000 UTC m=+224.029854448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.898045 4823 patch_prober.go:28] interesting pod/router-default-5444994796-lndmn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:56 crc kubenswrapper[4823]: [-]has-synced failed: reason withheld Mar 20 09:00:56 crc kubenswrapper[4823]: [+]process-running ok Mar 20 09:00:56 crc kubenswrapper[4823]: healthz check failed Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.898109 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lndmn" podUID="71d31547-2185-4113-a21e-2f0f2ae4423c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:56 crc kubenswrapper[4823]: I0320 09:00:56.945696 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:56 crc kubenswrapper[4823]: E0320 09:00:56.946020 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.446003655 +0000 UTC m=+224.131900029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.047664 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.048002 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.547990955 +0000 UTC m=+224.233887329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.149147 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.149527 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.649511842 +0000 UTC m=+224.335408216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.250837 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.251315 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.751299006 +0000 UTC m=+224.437195380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.261642 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" event={"ID":"3e566079-aa6d-4037-a58b-efc6721bc079","Type":"ContainerStarted","Data":"2e963e3505e6d5ac1a64b417db018d403cbe7227c71f886a1823e74a4eafa9e7"} Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.268424 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2jvr2" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.352702 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.354073 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.854028597 +0000 UTC m=+224.539924971 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.459336 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.459653 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:57.959636406 +0000 UTC m=+224.645532790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.476334 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7g599"] Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.477424 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.482761 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.495564 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7g599"] Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.570317 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.570717 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:58.070668954 +0000 UTC m=+224.756565328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.584849 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-utilities\") pod \"certified-operators-7g599\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.584961 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l482s\" (UniqueName: \"kubernetes.io/projected/95495063-7a29-4af1-a255-973a0129e289-kube-api-access-l482s\") pod \"certified-operators-7g599\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.585016 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.585055 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-catalog-content\") pod \"certified-operators-7g599\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.585587 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:58.085570225 +0000 UTC m=+224.771466599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.631928 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.632801 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.639319 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.639548 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.639911 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.640425 4823 ???:1] "http: TLS handshake error from 192.168.126.11:44142: no serving certificate available for the kubelet" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.643180 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p2lfv"] Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.643397 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" podUID="e0ccc792-7011-435d-934e-ad2d9d94ab90" containerName="controller-manager" containerID="cri-o://9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e" gracePeriod=30 Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.678134 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.687861 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.688044 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l482s\" (UniqueName: \"kubernetes.io/projected/95495063-7a29-4af1-a255-973a0129e289-kube-api-access-l482s\") pod \"certified-operators-7g599\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.688081 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-catalog-content\") pod \"certified-operators-7g599\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.688129 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bad5782e-a906-462c-a585-c3fa56499827-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"bad5782e-a906-462c-a585-c3fa56499827\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.688159 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-utilities\") pod \"certified-operators-7g599\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.688189 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bad5782e-a906-462c-a585-c3fa56499827-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"bad5782e-a906-462c-a585-c3fa56499827\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.688317 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:58.188301455 +0000 UTC m=+224.874197829 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.688986 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-catalog-content\") pod \"certified-operators-7g599\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.689195 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-utilities\") pod \"certified-operators-7g599\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.692078 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k9bv8"] Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.693102 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.709249 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.730291 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k9bv8"] Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.741221 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j"] Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.741468 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" podUID="ceb7506a-f654-4b9c-ba3d-d0fc6db798c0" containerName="route-controller-manager" containerID="cri-o://996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041" gracePeriod=30 Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.772198 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l482s\" (UniqueName: \"kubernetes.io/projected/95495063-7a29-4af1-a255-973a0129e289-kube-api-access-l482s\") pod \"certified-operators-7g599\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.782904 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.790582 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtcp6\" (UniqueName: \"kubernetes.io/projected/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-kube-api-access-dtcp6\") pod \"community-operators-k9bv8\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.790630 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bad5782e-a906-462c-a585-c3fa56499827-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"bad5782e-a906-462c-a585-c3fa56499827\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.790688 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-utilities\") pod \"community-operators-k9bv8\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.790716 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bad5782e-a906-462c-a585-c3fa56499827-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"bad5782e-a906-462c-a585-c3fa56499827\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.790845 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-catalog-content\") pod \"community-operators-k9bv8\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.790904 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.791239 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:58.291224221 +0000 UTC m=+224.977120595 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.791432 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bad5782e-a906-462c-a585-c3fa56499827-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"bad5782e-a906-462c-a585-c3fa56499827\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.814023 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.820713 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bad5782e-a906-462c-a585-c3fa56499827-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"bad5782e-a906-462c-a585-c3fa56499827\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.871966 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sj5ms"] Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.872486 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7" containerName="collect-profiles" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.872497 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7" containerName="collect-profiles" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.872587 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7" containerName="collect-profiles" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.873214 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.891587 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qfjh\" (UniqueName: \"kubernetes.io/projected/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-kube-api-access-9qfjh\") pod \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.891812 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.891853 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-config-volume\") pod \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.891902 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-secret-volume\") pod \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\" (UID: \"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7\") " Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.892064 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:58.392040159 +0000 UTC m=+225.077936533 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.892102 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtcp6\" (UniqueName: \"kubernetes.io/projected/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-kube-api-access-dtcp6\") pod \"community-operators-k9bv8\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.892209 4823 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.892972 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-config-volume" (OuterVolumeSpecName: "config-volume") pod "4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7" (UID: "4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.893220 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-utilities\") pod \"community-operators-k9bv8\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.892240 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-utilities\") pod \"community-operators-k9bv8\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.893474 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-catalog-content\") pod \"community-operators-k9bv8\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.893597 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.893798 4823 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.894525 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-catalog-content\") pod \"community-operators-k9bv8\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.896755 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:58.396722997 +0000 UTC m=+225.082619371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.897169 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7" (UID: "4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.900268 4823 patch_prober.go:28] interesting pod/router-default-5444994796-lndmn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:57 crc kubenswrapper[4823]: [-]has-synced failed: reason withheld Mar 20 09:00:57 crc kubenswrapper[4823]: [+]process-running ok Mar 20 09:00:57 crc kubenswrapper[4823]: healthz check failed Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.900343 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lndmn" podUID="71d31547-2185-4113-a21e-2f0f2ae4423c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.900806 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-kube-api-access-9qfjh" (OuterVolumeSpecName: "kube-api-access-9qfjh") pod "4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7" (UID: "4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7"). InnerVolumeSpecName "kube-api-access-9qfjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.914423 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtcp6\" (UniqueName: \"kubernetes.io/projected/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-kube-api-access-dtcp6\") pod \"community-operators-k9bv8\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.932487 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sj5ms"] Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.960041 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.995774 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.995972 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2df5\" (UniqueName: \"kubernetes.io/projected/4fafad3e-2369-4997-beb3-c73d884b2276-kube-api-access-c2df5\") pod \"certified-operators-sj5ms\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.996025 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-catalog-content\") pod \"certified-operators-sj5ms\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.996051 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-utilities\") pod \"certified-operators-sj5ms\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.996088 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qfjh\" (UniqueName: \"kubernetes.io/projected/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-kube-api-access-9qfjh\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:57 crc kubenswrapper[4823]: I0320 09:00:57.996098 4823 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:57 crc kubenswrapper[4823]: E0320 09:00:57.996173 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:58.496159366 +0000 UTC m=+225.182055740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.074835 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cbwgk"] Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.076249 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.088017 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cbwgk"] Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.101331 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-catalog-content\") pod \"certified-operators-sj5ms\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.101385 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-utilities\") pod \"certified-operators-sj5ms\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.101422 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.101464 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2df5\" (UniqueName: \"kubernetes.io/projected/4fafad3e-2369-4997-beb3-c73d884b2276-kube-api-access-c2df5\") pod \"certified-operators-sj5ms\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.101914 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-catalog-content\") pod \"certified-operators-sj5ms\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:00:58 crc kubenswrapper[4823]: E0320 09:00:58.102060 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:58.602048324 +0000 UTC m=+225.287944698 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.102087 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-utilities\") pod \"certified-operators-sj5ms\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.118815 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2df5\" (UniqueName: \"kubernetes.io/projected/4fafad3e-2369-4997-beb3-c73d884b2276-kube-api-access-c2df5\") pod \"certified-operators-sj5ms\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.201950 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:58 crc kubenswrapper[4823]: E0320 09:00:58.202103 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:58.70208067 +0000 UTC m=+225.387977044 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.202125 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-catalog-content\") pod \"community-operators-cbwgk\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.202163 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5fpf\" (UniqueName: \"kubernetes.io/projected/a0ddc452-8cdd-4451-b2a3-215c24347335-kube-api-access-n5fpf\") pod \"community-operators-cbwgk\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.202190 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.202275 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-utilities\") pod \"community-operators-cbwgk\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:00:58 crc kubenswrapper[4823]: E0320 09:00:58.202431 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:58.702419719 +0000 UTC m=+225.388316143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwrsf" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.205843 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.212767 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.220531 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.223856 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.282773 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.282773 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8" event={"ID":"4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7","Type":"ContainerDied","Data":"a7b50c5fd9f539a08f3a442445fbae533a4aecc59fbcc67adeeef7988a164e8d"} Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.282908 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7b50c5fd9f539a08f3a442445fbae533a4aecc59fbcc67adeeef7988a164e8d" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.284726 4823 generic.go:334] "Generic (PLEG): container finished" podID="ceb7506a-f654-4b9c-ba3d-d0fc6db798c0" containerID="996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041" exitCode=0 Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.284783 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" event={"ID":"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0","Type":"ContainerDied","Data":"996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041"} Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.284807 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" event={"ID":"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0","Type":"ContainerDied","Data":"89793889787aad51372aaaf0da12df7e5d2e7ba2309e4b4eb3418355516bb1e7"} Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.284823 4823 scope.go:117] "RemoveContainer" containerID="996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.284915 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.284918 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.303500 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-config\") pod \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.303560 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k94q\" (UniqueName: \"kubernetes.io/projected/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-kube-api-access-2k94q\") pod \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.303710 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.303763 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-proxy-ca-bundles\") pod \"e0ccc792-7011-435d-934e-ad2d9d94ab90\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.303789 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-serving-cert\") pod \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.303824 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl9dg\" (UniqueName: \"kubernetes.io/projected/e0ccc792-7011-435d-934e-ad2d9d94ab90-kube-api-access-jl9dg\") pod \"e0ccc792-7011-435d-934e-ad2d9d94ab90\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.303868 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0ccc792-7011-435d-934e-ad2d9d94ab90-serving-cert\") pod \"e0ccc792-7011-435d-934e-ad2d9d94ab90\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.303902 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-client-ca\") pod \"e0ccc792-7011-435d-934e-ad2d9d94ab90\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.303922 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-config\") pod \"e0ccc792-7011-435d-934e-ad2d9d94ab90\" (UID: \"e0ccc792-7011-435d-934e-ad2d9d94ab90\") " Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.303944 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-client-ca\") pod \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\" (UID: \"ceb7506a-f654-4b9c-ba3d-d0fc6db798c0\") " Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.304267 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-catalog-content\") pod \"community-operators-cbwgk\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.304304 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5fpf\" (UniqueName: \"kubernetes.io/projected/a0ddc452-8cdd-4451-b2a3-215c24347335-kube-api-access-n5fpf\") pod \"community-operators-cbwgk\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.304345 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-utilities\") pod \"community-operators-cbwgk\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:00:58 crc kubenswrapper[4823]: E0320 09:00:58.306775 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:58.806711752 +0000 UTC m=+225.492608126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.307072 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-config" (OuterVolumeSpecName: "config") pod "ceb7506a-f654-4b9c-ba3d-d0fc6db798c0" (UID: "ceb7506a-f654-4b9c-ba3d-d0fc6db798c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.307595 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-catalog-content\") pod \"community-operators-cbwgk\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.308511 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-client-ca" (OuterVolumeSpecName: "client-ca") pod "ceb7506a-f654-4b9c-ba3d-d0fc6db798c0" (UID: "ceb7506a-f654-4b9c-ba3d-d0fc6db798c0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.310075 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-client-ca" (OuterVolumeSpecName: "client-ca") pod "e0ccc792-7011-435d-934e-ad2d9d94ab90" (UID: "e0ccc792-7011-435d-934e-ad2d9d94ab90"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.310406 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e0ccc792-7011-435d-934e-ad2d9d94ab90" (UID: "e0ccc792-7011-435d-934e-ad2d9d94ab90"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.310440 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" event={"ID":"3e566079-aa6d-4037-a58b-efc6721bc079","Type":"ContainerStarted","Data":"24ed4d96549b61fd19e091899069c92bef4ad37543381c00f5cea7e4932ec572"} Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.310481 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" event={"ID":"3e566079-aa6d-4037-a58b-efc6721bc079","Type":"ContainerStarted","Data":"31d67191962e686865b6fba631bd33466c79ad9972d4c3589de142fc0a72694d"} Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.310514 4823 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-20T09:00:57.892292525Z","Handler":null,"Name":""} Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.311200 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-config" (OuterVolumeSpecName: "config") pod "e0ccc792-7011-435d-934e-ad2d9d94ab90" (UID: "e0ccc792-7011-435d-934e-ad2d9d94ab90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:58 crc kubenswrapper[4823]: W0320 09:00:58.313443 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbad5782e_a906_462c_a585_c3fa56499827.slice/crio-17f9cd9539a4720c5f8c2001629c9736608fac226c3957bc2408fc765a187ef7 WatchSource:0}: Error finding container 17f9cd9539a4720c5f8c2001629c9736608fac226c3957bc2408fc765a187ef7: Status 404 returned error can't find the container with id 17f9cd9539a4720c5f8c2001629c9736608fac226c3957bc2408fc765a187ef7 Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.314447 4823 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.314977 4823 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.315696 4823 generic.go:334] "Generic (PLEG): container finished" podID="e0ccc792-7011-435d-934e-ad2d9d94ab90" containerID="9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e" exitCode=0 Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.316195 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.316273 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" event={"ID":"e0ccc792-7011-435d-934e-ad2d9d94ab90","Type":"ContainerDied","Data":"9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e"} Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.316326 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ceb7506a-f654-4b9c-ba3d-d0fc6db798c0" (UID: "ceb7506a-f654-4b9c-ba3d-d0fc6db798c0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.316362 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-p2lfv" event={"ID":"e0ccc792-7011-435d-934e-ad2d9d94ab90","Type":"ContainerDied","Data":"dcc68970de824bc41ee6f33879a2a204dd30f70134a32531d3d6f788539b9bbf"} Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.317768 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-kube-api-access-2k94q" (OuterVolumeSpecName: "kube-api-access-2k94q") pod "ceb7506a-f654-4b9c-ba3d-d0fc6db798c0" (UID: "ceb7506a-f654-4b9c-ba3d-d0fc6db798c0"). InnerVolumeSpecName "kube-api-access-2k94q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.321418 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0ccc792-7011-435d-934e-ad2d9d94ab90-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e0ccc792-7011-435d-934e-ad2d9d94ab90" (UID: "e0ccc792-7011-435d-934e-ad2d9d94ab90"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.322515 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ccc792-7011-435d-934e-ad2d9d94ab90-kube-api-access-jl9dg" (OuterVolumeSpecName: "kube-api-access-jl9dg") pod "e0ccc792-7011-435d-934e-ad2d9d94ab90" (UID: "e0ccc792-7011-435d-934e-ad2d9d94ab90"). InnerVolumeSpecName "kube-api-access-jl9dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.326470 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-utilities\") pod \"community-operators-cbwgk\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.329400 4823 scope.go:117] "RemoveContainer" containerID="996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.329845 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-6cfw8" podStartSLOduration=11.329831499 podStartE2EDuration="11.329831499s" podCreationTimestamp="2026-03-20 09:00:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:58.328162793 +0000 UTC m=+225.014059167" watchObservedRunningTime="2026-03-20 09:00:58.329831499 +0000 UTC m=+225.015727873" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.331600 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5fpf\" (UniqueName: \"kubernetes.io/projected/a0ddc452-8cdd-4451-b2a3-215c24347335-kube-api-access-n5fpf\") pod \"community-operators-cbwgk\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:00:58 crc kubenswrapper[4823]: E0320 09:00:58.333080 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041\": container with ID starting with 996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041 not found: ID does not exist" containerID="996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.333135 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041"} err="failed to get container status \"996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041\": rpc error: code = NotFound desc = could not find container \"996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041\": container with ID starting with 996b77526ecfb0afb859a15775a802e8bf483ca781c35a83bbbf601ce8d5e041 not found: ID does not exist" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.333163 4823 scope.go:117] "RemoveContainer" containerID="9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.359622 4823 scope.go:117] "RemoveContainer" containerID="9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e" Mar 20 09:00:58 crc kubenswrapper[4823]: E0320 09:00:58.360378 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e\": container with ID starting with 9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e not found: ID does not exist" containerID="9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.360437 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e"} err="failed to get container status \"9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e\": rpc error: code = NotFound desc = could not find container \"9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e\": container with ID starting with 9ae40af99e8bd21de0c66ef1844d3bb41a81513b341333aed9ab5fad4b87d58e not found: ID does not exist" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.399118 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.408937 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.409084 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.409096 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k94q\" (UniqueName: \"kubernetes.io/projected/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-kube-api-access-2k94q\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.409107 4823 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.409115 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.409124 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl9dg\" (UniqueName: \"kubernetes.io/projected/e0ccc792-7011-435d-934e-ad2d9d94ab90-kube-api-access-jl9dg\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.409133 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0ccc792-7011-435d-934e-ad2d9d94ab90-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.409140 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.409149 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0ccc792-7011-435d-934e-ad2d9d94ab90-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.409156 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.415901 4823 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.416349 4823 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.423948 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7g599"] Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.448839 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwrsf\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.479285 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.512395 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.512828 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sj5ms"] Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.535288 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.572088 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k9bv8"] Mar 20 09:00:58 crc kubenswrapper[4823]: W0320 09:00:58.574561 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fafad3e_2369_4997_beb3_c73d884b2276.slice/crio-2d02cbcbf3b3f8cad7228ae4c76ce24d4928bd90d48d49b900d0e907150dab46 WatchSource:0}: Error finding container 2d02cbcbf3b3f8cad7228ae4c76ce24d4928bd90d48d49b900d0e907150dab46: Status 404 returned error can't find the container with id 2d02cbcbf3b3f8cad7228ae4c76ce24d4928bd90d48d49b900d0e907150dab46 Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.635320 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j"] Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.643222 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dm29j"] Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.670854 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p2lfv"] Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.675120 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-p2lfv"] Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.676874 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cbwgk"] Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.833618 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwrsf"] Mar 20 09:00:58 crc kubenswrapper[4823]: W0320 09:00:58.852235 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9000667_ed2e_4fea_b03f_c9760e59a604.slice/crio-fb4a34145e231b8f132c9acfcac8fd1d470eec42959da4548ad7ea324f91e57c WatchSource:0}: Error finding container fb4a34145e231b8f132c9acfcac8fd1d470eec42959da4548ad7ea324f91e57c: Status 404 returned error can't find the container with id fb4a34145e231b8f132c9acfcac8fd1d470eec42959da4548ad7ea324f91e57c Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.899453 4823 patch_prober.go:28] interesting pod/router-default-5444994796-lndmn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:58 crc kubenswrapper[4823]: [-]has-synced failed: reason withheld Mar 20 09:00:58 crc kubenswrapper[4823]: [+]process-running ok Mar 20 09:00:58 crc kubenswrapper[4823]: healthz check failed Mar 20 09:00:58 crc kubenswrapper[4823]: I0320 09:00:58.899516 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lndmn" podUID="71d31547-2185-4113-a21e-2f0f2ae4423c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.121523 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.121880 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.126452 4823 patch_prober.go:28] interesting pod/console-f9d7485db-lz4n9 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.126582 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-lz4n9" podUID="e104c875-7420-45b3-b675-a63b92c454e4" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.137855 4823 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.164413 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 09:00:59 crc kubenswrapper[4823]: E0320 09:00:59.164631 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ccc792-7011-435d-934e-ad2d9d94ab90" containerName="controller-manager" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.164643 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ccc792-7011-435d-934e-ad2d9d94ab90" containerName="controller-manager" Mar 20 09:00:59 crc kubenswrapper[4823]: E0320 09:00:59.164650 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb7506a-f654-4b9c-ba3d-d0fc6db798c0" containerName="route-controller-manager" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.164658 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb7506a-f654-4b9c-ba3d-d0fc6db798c0" containerName="route-controller-manager" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.164776 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceb7506a-f654-4b9c-ba3d-d0fc6db798c0" containerName="route-controller-manager" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.164790 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ccc792-7011-435d-934e-ad2d9d94ab90" containerName="controller-manager" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.165148 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.175292 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.177951 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.178168 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.222238 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e3b2ffe1-5822-46ce-b1ee-23eb79596b97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.222380 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e3b2ffe1-5822-46ce-b1ee-23eb79596b97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.323094 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e3b2ffe1-5822-46ce-b1ee-23eb79596b97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.323172 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e3b2ffe1-5822-46ce-b1ee-23eb79596b97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.323238 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e3b2ffe1-5822-46ce-b1ee-23eb79596b97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.325383 4823 generic.go:334] "Generic (PLEG): container finished" podID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" containerID="58ba417b79c9ab68df891927d11ee0bdf0405d4689ddeee8346c61122c0fd079" exitCode=0 Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.325468 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9bv8" event={"ID":"f9cc626b-cf2b-407f-ac25-c9ce01e0660d","Type":"ContainerDied","Data":"58ba417b79c9ab68df891927d11ee0bdf0405d4689ddeee8346c61122c0fd079"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.325520 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9bv8" event={"ID":"f9cc626b-cf2b-407f-ac25-c9ce01e0660d","Type":"ContainerStarted","Data":"4d2487d8c4e1a7b21639e871db613c44418e15e647f5e5f4895aa1542523a8c0"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.328154 4823 generic.go:334] "Generic (PLEG): container finished" podID="a0ddc452-8cdd-4451-b2a3-215c24347335" containerID="a8b057a779abaefe8364a70d4ffea1cb0723b145bb437daef989f1ea146f40ad" exitCode=0 Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.328236 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbwgk" event={"ID":"a0ddc452-8cdd-4451-b2a3-215c24347335","Type":"ContainerDied","Data":"a8b057a779abaefe8364a70d4ffea1cb0723b145bb437daef989f1ea146f40ad"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.328276 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbwgk" event={"ID":"a0ddc452-8cdd-4451-b2a3-215c24347335","Type":"ContainerStarted","Data":"59ee1ad4b8cd4342b517e4dc68d5641250b4e61319d47c4da45dfb4e98b755a7"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.357949 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e3b2ffe1-5822-46ce-b1ee-23eb79596b97\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.361173 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bad5782e-a906-462c-a585-c3fa56499827","Type":"ContainerStarted","Data":"dd8a5175a5a013251f036cf979ecd430ba7401c4f18aab7e7358493156a2320e"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.361213 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bad5782e-a906-462c-a585-c3fa56499827","Type":"ContainerStarted","Data":"17f9cd9539a4720c5f8c2001629c9736608fac226c3957bc2408fc765a187ef7"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.395585 4823 generic.go:334] "Generic (PLEG): container finished" podID="4fafad3e-2369-4997-beb3-c73d884b2276" containerID="47dc9a523f536c8f14d995f072b38e6c61ee7c4bd45af89cd627bab79f7bd90a" exitCode=0 Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.395763 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj5ms" event={"ID":"4fafad3e-2369-4997-beb3-c73d884b2276","Type":"ContainerDied","Data":"47dc9a523f536c8f14d995f072b38e6c61ee7c4bd45af89cd627bab79f7bd90a"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.395797 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj5ms" event={"ID":"4fafad3e-2369-4997-beb3-c73d884b2276","Type":"ContainerStarted","Data":"2d02cbcbf3b3f8cad7228ae4c76ce24d4928bd90d48d49b900d0e907150dab46"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.400167 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-754fdbdd85-xt9c4"] Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.401179 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.411337 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.413165 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.413141823 podStartE2EDuration="2.413141823s" podCreationTimestamp="2026-03-20 09:00:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:59.398455328 +0000 UTC m=+226.084351702" watchObservedRunningTime="2026-03-20 09:00:59.413141823 +0000 UTC m=+226.099038197" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.414808 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj"] Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.415760 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.429467 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.431516 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.431768 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.432115 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.432635 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.433383 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" event={"ID":"e9000667-ed2e-4fea-b03f-c9760e59a604","Type":"ContainerStarted","Data":"d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.433481 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" event={"ID":"e9000667-ed2e-4fea-b03f-c9760e59a604","Type":"ContainerStarted","Data":"fb4a34145e231b8f132c9acfcac8fd1d470eec42959da4548ad7ea324f91e57c"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.433650 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.435484 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.436444 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.436096 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.436843 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.439906 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.440164 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.440332 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.443022 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-754fdbdd85-xt9c4"] Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.447380 4823 generic.go:334] "Generic (PLEG): container finished" podID="95495063-7a29-4af1-a255-973a0129e289" containerID="a54dd9c64820ad8e9843816ed6926c74b7e3d824b002b853cd38c2200d1f9bbc" exitCode=0 Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.448278 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g599" event={"ID":"95495063-7a29-4af1-a255-973a0129e289","Type":"ContainerDied","Data":"a54dd9c64820ad8e9843816ed6926c74b7e3d824b002b853cd38c2200d1f9bbc"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.448406 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj"] Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.448502 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g599" event={"ID":"95495063-7a29-4af1-a255-973a0129e289","Type":"ContainerStarted","Data":"f6e1dbc892b868ff239a259c65c96139dd9b5a41762bae781e1fe154713c4cf2"} Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.491732 4823 patch_prober.go:28] interesting pod/downloads-7954f5f757-grjk6 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.492110 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-grjk6" podUID="98f13ee0-723b-4ef5-851b-2665009a9d47" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.491757 4823 patch_prober.go:28] interesting pod/downloads-7954f5f757-grjk6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.492363 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grjk6" podUID="98f13ee0-723b-4ef5-851b-2665009a9d47" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.500231 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6m5xl"] Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.501409 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.505769 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6m5xl"] Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.510484 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.514088 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.526543 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-config\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.526696 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-client-ca\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.526806 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa5a375e-efce-4bf6-a949-37de5cd53861-serving-cert\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.526894 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwcnz\" (UniqueName: \"kubernetes.io/projected/fa5a375e-efce-4bf6-a949-37de5cd53861-kube-api-access-nwcnz\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.526976 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-proxy-ca-bundles\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.527047 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-config\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.527160 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-serving-cert\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.527243 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qptv8\" (UniqueName: \"kubernetes.io/projected/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-kube-api-access-qptv8\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.527329 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-client-ca\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.553640 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" podStartSLOduration=184.553603913 podStartE2EDuration="3m4.553603913s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:59.544924994 +0000 UTC m=+226.230821368" watchObservedRunningTime="2026-03-20 09:00:59.553603913 +0000 UTC m=+226.239500287" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629673 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-serving-cert\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629717 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qptv8\" (UniqueName: \"kubernetes.io/projected/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-kube-api-access-qptv8\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629760 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-utilities\") pod \"redhat-marketplace-6m5xl\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629779 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-client-ca\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629797 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-config\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629830 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-client-ca\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629860 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa5a375e-efce-4bf6-a949-37de5cd53861-serving-cert\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629879 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzkf2\" (UniqueName: \"kubernetes.io/projected/2cee23f5-1268-4adf-8466-da5931de9825-kube-api-access-fzkf2\") pod \"redhat-marketplace-6m5xl\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629899 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwcnz\" (UniqueName: \"kubernetes.io/projected/fa5a375e-efce-4bf6-a949-37de5cd53861-kube-api-access-nwcnz\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629918 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-proxy-ca-bundles\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629940 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-config\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.629956 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-catalog-content\") pod \"redhat-marketplace-6m5xl\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.635028 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-client-ca\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.635813 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-config\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.636059 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-proxy-ca-bundles\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.636931 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-config\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.636952 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-client-ca\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.643251 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa5a375e-efce-4bf6-a949-37de5cd53861-serving-cert\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.652606 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-serving-cert\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.653511 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwcnz\" (UniqueName: \"kubernetes.io/projected/fa5a375e-efce-4bf6-a949-37de5cd53861-kube-api-access-nwcnz\") pod \"route-controller-manager-797bf5c8dc-clhqj\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.653623 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qptv8\" (UniqueName: \"kubernetes.io/projected/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-kube-api-access-qptv8\") pod \"controller-manager-754fdbdd85-xt9c4\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.698171 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.698396 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.707027 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.731530 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzkf2\" (UniqueName: \"kubernetes.io/projected/2cee23f5-1268-4adf-8466-da5931de9825-kube-api-access-fzkf2\") pod \"redhat-marketplace-6m5xl\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.731578 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-catalog-content\") pod \"redhat-marketplace-6m5xl\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.731623 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-utilities\") pod \"redhat-marketplace-6m5xl\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.732054 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-utilities\") pod \"redhat-marketplace-6m5xl\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.732489 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-catalog-content\") pod \"redhat-marketplace-6m5xl\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.751904 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.753167 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzkf2\" (UniqueName: \"kubernetes.io/projected/2cee23f5-1268-4adf-8466-da5931de9825-kube-api-access-fzkf2\") pod \"redhat-marketplace-6m5xl\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.786037 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.797965 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.818774 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.862678 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-425tz"] Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.863783 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.873904 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-425tz"] Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.895980 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.910240 4823 patch_prober.go:28] interesting pod/router-default-5444994796-lndmn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:59 crc kubenswrapper[4823]: [-]has-synced failed: reason withheld Mar 20 09:00:59 crc kubenswrapper[4823]: [+]process-running ok Mar 20 09:00:59 crc kubenswrapper[4823]: healthz check failed Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.910329 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lndmn" podUID="71d31547-2185-4113-a21e-2f0f2ae4423c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.935709 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-utilities\") pod \"redhat-marketplace-425tz\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.935786 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-catalog-content\") pod \"redhat-marketplace-425tz\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.935922 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7772\" (UniqueName: \"kubernetes.io/projected/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-kube-api-access-t7772\") pod \"redhat-marketplace-425tz\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:00:59 crc kubenswrapper[4823]: I0320 09:00:59.940406 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-qsxgq" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.040134 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7772\" (UniqueName: \"kubernetes.io/projected/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-kube-api-access-t7772\") pod \"redhat-marketplace-425tz\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.040873 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-utilities\") pod \"redhat-marketplace-425tz\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.040899 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-catalog-content\") pod \"redhat-marketplace-425tz\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.054133 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-catalog-content\") pod \"redhat-marketplace-425tz\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.054203 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-utilities\") pod \"redhat-marketplace-425tz\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.063683 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7772\" (UniqueName: \"kubernetes.io/projected/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-kube-api-access-t7772\") pod \"redhat-marketplace-425tz\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.115503 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.116179 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceb7506a-f654-4b9c-ba3d-d0fc6db798c0" path="/var/lib/kubelet/pods/ceb7506a-f654-4b9c-ba3d-d0fc6db798c0/volumes" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.116697 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0ccc792-7011-435d-934e-ad2d9d94ab90" path="/var/lib/kubelet/pods/e0ccc792-7011-435d-934e-ad2d9d94ab90/volumes" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.210865 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.234564 4823 ???:1] "http: TLS handshake error from 192.168.126.11:44154: no serving certificate available for the kubelet" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.267694 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-754fdbdd85-xt9c4"] Mar 20 09:01:00 crc kubenswrapper[4823]: W0320 09:01:00.300586 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21d3c4ae_d6f7_421b_bf3d_f905a74b3db6.slice/crio-0dbc7949e2ef9cf323cfe05579325d13e82b32b0cd80544608b8e5ec3a4e7ce5 WatchSource:0}: Error finding container 0dbc7949e2ef9cf323cfe05579325d13e82b32b0cd80544608b8e5ec3a4e7ce5: Status 404 returned error can't find the container with id 0dbc7949e2ef9cf323cfe05579325d13e82b32b0cd80544608b8e5ec3a4e7ce5 Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.373804 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6m5xl"] Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.376527 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj"] Mar 20 09:01:00 crc kubenswrapper[4823]: W0320 09:01:00.391827 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cee23f5_1268_4adf_8466_da5931de9825.slice/crio-2fd3f02fd2f3d952dc91769865ba11a2ae3209c1459e4102154dd29a48b462f1 WatchSource:0}: Error finding container 2fd3f02fd2f3d952dc91769865ba11a2ae3209c1459e4102154dd29a48b462f1: Status 404 returned error can't find the container with id 2fd3f02fd2f3d952dc91769865ba11a2ae3209c1459e4102154dd29a48b462f1 Mar 20 09:01:00 crc kubenswrapper[4823]: W0320 09:01:00.404525 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa5a375e_efce_4bf6_a949_37de5cd53861.slice/crio-f5c0f9108bb074c32d438269fbe7f0c07834317dba5659ddadaaa892e8f57be5 WatchSource:0}: Error finding container f5c0f9108bb074c32d438269fbe7f0c07834317dba5659ddadaaa892e8f57be5: Status 404 returned error can't find the container with id f5c0f9108bb074c32d438269fbe7f0c07834317dba5659ddadaaa892e8f57be5 Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.481814 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e3b2ffe1-5822-46ce-b1ee-23eb79596b97","Type":"ContainerStarted","Data":"df3acce9151687097c3ca1446b550f7beb540d7c52b6a2f253f84f37da7304cc"} Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.481895 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e3b2ffe1-5822-46ce-b1ee-23eb79596b97","Type":"ContainerStarted","Data":"c1dac87eb5431c6f5418e041e7aa368f1cc2f8dd20efcaa61f1552479d966f67"} Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.484020 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-425tz"] Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.489257 4823 generic.go:334] "Generic (PLEG): container finished" podID="bad5782e-a906-462c-a585-c3fa56499827" containerID="dd8a5175a5a013251f036cf979ecd430ba7401c4f18aab7e7358493156a2320e" exitCode=0 Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.489299 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bad5782e-a906-462c-a585-c3fa56499827","Type":"ContainerDied","Data":"dd8a5175a5a013251f036cf979ecd430ba7401c4f18aab7e7358493156a2320e"} Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.502327 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" event={"ID":"fa5a375e-efce-4bf6-a949-37de5cd53861","Type":"ContainerStarted","Data":"f5c0f9108bb074c32d438269fbe7f0c07834317dba5659ddadaaa892e8f57be5"} Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.503094 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.50305029 podStartE2EDuration="1.50305029s" podCreationTimestamp="2026-03-20 09:00:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:00.499327267 +0000 UTC m=+227.185223651" watchObservedRunningTime="2026-03-20 09:01:00.50305029 +0000 UTC m=+227.188946664" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.511704 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" event={"ID":"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6","Type":"ContainerStarted","Data":"0dbc7949e2ef9cf323cfe05579325d13e82b32b0cd80544608b8e5ec3a4e7ce5"} Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.521934 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m5xl" event={"ID":"2cee23f5-1268-4adf-8466-da5931de9825","Type":"ContainerStarted","Data":"2fd3f02fd2f3d952dc91769865ba11a2ae3209c1459e4102154dd29a48b462f1"} Mar 20 09:01:00 crc kubenswrapper[4823]: W0320 09:01:00.522834 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86c6eee3_c57f_41af_8bf5_73f1c15ca2fe.slice/crio-229a0693bfd2db0add9998f62ebf35d8efbfc8c149524af77853ae2ca42c766a WatchSource:0}: Error finding container 229a0693bfd2db0add9998f62ebf35d8efbfc8c149524af77853ae2ca42c766a: Status 404 returned error can't find the container with id 229a0693bfd2db0add9998f62ebf35d8efbfc8c149524af77853ae2ca42c766a Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.527591 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-sswrk" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.673587 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vbgxd"] Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.675071 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.679280 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.690906 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vbgxd"] Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.756668 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-utilities\") pod \"redhat-operators-vbgxd\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.756760 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc56b\" (UniqueName: \"kubernetes.io/projected/694805b6-f8a1-4277-8132-100503fdc1a8-kube-api-access-pc56b\") pod \"redhat-operators-vbgxd\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.756793 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-catalog-content\") pod \"redhat-operators-vbgxd\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.849163 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cgkx8" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.858435 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc56b\" (UniqueName: \"kubernetes.io/projected/694805b6-f8a1-4277-8132-100503fdc1a8-kube-api-access-pc56b\") pod \"redhat-operators-vbgxd\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.858525 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-catalog-content\") pod \"redhat-operators-vbgxd\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.858641 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-utilities\") pod \"redhat-operators-vbgxd\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.859352 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-utilities\") pod \"redhat-operators-vbgxd\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.860121 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-catalog-content\") pod \"redhat-operators-vbgxd\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.902586 4823 patch_prober.go:28] interesting pod/router-default-5444994796-lndmn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:01:00 crc kubenswrapper[4823]: [-]has-synced failed: reason withheld Mar 20 09:01:00 crc kubenswrapper[4823]: [+]process-running ok Mar 20 09:01:00 crc kubenswrapper[4823]: healthz check failed Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.902980 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lndmn" podUID="71d31547-2185-4113-a21e-2f0f2ae4423c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:01:00 crc kubenswrapper[4823]: I0320 09:01:00.909845 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc56b\" (UniqueName: \"kubernetes.io/projected/694805b6-f8a1-4277-8132-100503fdc1a8-kube-api-access-pc56b\") pod \"redhat-operators-vbgxd\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.057964 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.062756 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b482j"] Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.064014 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.086630 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b482j"] Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.168850 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pswd\" (UniqueName: \"kubernetes.io/projected/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-kube-api-access-4pswd\") pod \"redhat-operators-b482j\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.168980 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-utilities\") pod \"redhat-operators-b482j\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.169140 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-catalog-content\") pod \"redhat-operators-b482j\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.286144 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-utilities\") pod \"redhat-operators-b482j\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.292318 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-utilities\") pod \"redhat-operators-b482j\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.292535 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-catalog-content\") pod \"redhat-operators-b482j\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.292717 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pswd\" (UniqueName: \"kubernetes.io/projected/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-kube-api-access-4pswd\") pod \"redhat-operators-b482j\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.293811 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-catalog-content\") pod \"redhat-operators-b482j\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.317145 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pswd\" (UniqueName: \"kubernetes.io/projected/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-kube-api-access-4pswd\") pod \"redhat-operators-b482j\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.462552 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vbgxd"] Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.477010 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:01 crc kubenswrapper[4823]: W0320 09:01:01.556290 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod694805b6_f8a1_4277_8132_100503fdc1a8.slice/crio-b920167dcb12e251075f504088e9c1cd6db5443d9bf8f5370b4c75c86439d0db WatchSource:0}: Error finding container b920167dcb12e251075f504088e9c1cd6db5443d9bf8f5370b4c75c86439d0db: Status 404 returned error can't find the container with id b920167dcb12e251075f504088e9c1cd6db5443d9bf8f5370b4c75c86439d0db Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.580096 4823 generic.go:334] "Generic (PLEG): container finished" podID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" containerID="c2f21152a51a4ef4355aaa4144fd73da9cf09e78184cf4f9e558105251a18be4" exitCode=0 Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.580187 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-425tz" event={"ID":"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe","Type":"ContainerDied","Data":"c2f21152a51a4ef4355aaa4144fd73da9cf09e78184cf4f9e558105251a18be4"} Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.580215 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-425tz" event={"ID":"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe","Type":"ContainerStarted","Data":"229a0693bfd2db0add9998f62ebf35d8efbfc8c149524af77853ae2ca42c766a"} Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.594950 4823 generic.go:334] "Generic (PLEG): container finished" podID="2cee23f5-1268-4adf-8466-da5931de9825" containerID="b9674954a944094d62dfc8be3a4182df4ae1cafc9d7ee9c762ad9a7baacd9f3d" exitCode=0 Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.595031 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m5xl" event={"ID":"2cee23f5-1268-4adf-8466-da5931de9825","Type":"ContainerDied","Data":"b9674954a944094d62dfc8be3a4182df4ae1cafc9d7ee9c762ad9a7baacd9f3d"} Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.633624 4823 generic.go:334] "Generic (PLEG): container finished" podID="e3b2ffe1-5822-46ce-b1ee-23eb79596b97" containerID="df3acce9151687097c3ca1446b550f7beb540d7c52b6a2f253f84f37da7304cc" exitCode=0 Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.633669 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e3b2ffe1-5822-46ce-b1ee-23eb79596b97","Type":"ContainerDied","Data":"df3acce9151687097c3ca1446b550f7beb540d7c52b6a2f253f84f37da7304cc"} Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.641407 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" event={"ID":"fa5a375e-efce-4bf6-a949-37de5cd53861","Type":"ContainerStarted","Data":"4a00ff2c23855f1d20e07bace0ab606b9b0a76f3906354f34917014e0a5665ea"} Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.641996 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.648008 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" event={"ID":"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6","Type":"ContainerStarted","Data":"efd92f99adb9a9ac56756c38189daf7dab1234a10ca4b0ec1bad041e78d49b44"} Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.648055 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.654555 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.667286 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.684124 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" podStartSLOduration=4.684108747 podStartE2EDuration="4.684108747s" podCreationTimestamp="2026-03-20 09:00:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:01.67877241 +0000 UTC m=+228.364668804" watchObservedRunningTime="2026-03-20 09:01:01.684108747 +0000 UTC m=+228.370005121" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.704018 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" podStartSLOduration=4.704002265 podStartE2EDuration="4.704002265s" podCreationTimestamp="2026-03-20 09:00:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:01.701890937 +0000 UTC m=+228.387787301" watchObservedRunningTime="2026-03-20 09:01:01.704002265 +0000 UTC m=+228.389898639" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.911364 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:01:01 crc kubenswrapper[4823]: I0320 09:01:01.925075 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-lndmn" Mar 20 09:01:02 crc kubenswrapper[4823]: I0320 09:01:02.298335 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b482j"] Mar 20 09:01:02 crc kubenswrapper[4823]: I0320 09:01:02.674308 4823 generic.go:334] "Generic (PLEG): container finished" podID="694805b6-f8a1-4277-8132-100503fdc1a8" containerID="fa834b17f4012d635f99377eb4f8ace23e290542189ef037ee00b85fe968ab1f" exitCode=0 Mar 20 09:01:02 crc kubenswrapper[4823]: I0320 09:01:02.675701 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbgxd" event={"ID":"694805b6-f8a1-4277-8132-100503fdc1a8","Type":"ContainerDied","Data":"fa834b17f4012d635f99377eb4f8ace23e290542189ef037ee00b85fe968ab1f"} Mar 20 09:01:02 crc kubenswrapper[4823]: I0320 09:01:02.675761 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbgxd" event={"ID":"694805b6-f8a1-4277-8132-100503fdc1a8","Type":"ContainerStarted","Data":"b920167dcb12e251075f504088e9c1cd6db5443d9bf8f5370b4c75c86439d0db"} Mar 20 09:01:05 crc kubenswrapper[4823]: I0320 09:01:05.377019 4823 ???:1] "http: TLS handshake error from 192.168.126.11:45786: no serving certificate available for the kubelet" Mar 20 09:01:05 crc kubenswrapper[4823]: I0320 09:01:05.533861 4823 ???:1] "http: TLS handshake error from 192.168.126.11:45802: no serving certificate available for the kubelet" Mar 20 09:01:05 crc kubenswrapper[4823]: I0320 09:01:05.816759 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-76jgq" Mar 20 09:01:08 crc kubenswrapper[4823]: W0320 09:01:08.628755 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde9e8a09_5d12_42e9_aba9_7fb5ec3cc6fc.slice/crio-e44db6133223dc93f4e83f53f1bf50ed6250d5f417bac0b1fbf2c6c5aec158e2 WatchSource:0}: Error finding container e44db6133223dc93f4e83f53f1bf50ed6250d5f417bac0b1fbf2c6c5aec158e2: Status 404 returned error can't find the container with id e44db6133223dc93f4e83f53f1bf50ed6250d5f417bac0b1fbf2c6c5aec158e2 Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.666294 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.673919 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.728222 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kubelet-dir\") pod \"e3b2ffe1-5822-46ce-b1ee-23eb79596b97\" (UID: \"e3b2ffe1-5822-46ce-b1ee-23eb79596b97\") " Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.728554 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bad5782e-a906-462c-a585-c3fa56499827-kubelet-dir\") pod \"bad5782e-a906-462c-a585-c3fa56499827\" (UID: \"bad5782e-a906-462c-a585-c3fa56499827\") " Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.728599 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bad5782e-a906-462c-a585-c3fa56499827-kube-api-access\") pod \"bad5782e-a906-462c-a585-c3fa56499827\" (UID: \"bad5782e-a906-462c-a585-c3fa56499827\") " Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.728705 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kube-api-access\") pod \"e3b2ffe1-5822-46ce-b1ee-23eb79596b97\" (UID: \"e3b2ffe1-5822-46ce-b1ee-23eb79596b97\") " Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.729862 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e3b2ffe1-5822-46ce-b1ee-23eb79596b97" (UID: "e3b2ffe1-5822-46ce-b1ee-23eb79596b97"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.730308 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bad5782e-a906-462c-a585-c3fa56499827-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bad5782e-a906-462c-a585-c3fa56499827" (UID: "bad5782e-a906-462c-a585-c3fa56499827"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.752923 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e3b2ffe1-5822-46ce-b1ee-23eb79596b97" (UID: "e3b2ffe1-5822-46ce-b1ee-23eb79596b97"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.754990 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bad5782e-a906-462c-a585-c3fa56499827-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bad5782e-a906-462c-a585-c3fa56499827" (UID: "bad5782e-a906-462c-a585-c3fa56499827"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.778427 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.779330 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e3b2ffe1-5822-46ce-b1ee-23eb79596b97","Type":"ContainerDied","Data":"c1dac87eb5431c6f5418e041e7aa368f1cc2f8dd20efcaa61f1552479d966f67"} Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.779389 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1dac87eb5431c6f5418e041e7aa368f1cc2f8dd20efcaa61f1552479d966f67" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.785508 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"bad5782e-a906-462c-a585-c3fa56499827","Type":"ContainerDied","Data":"17f9cd9539a4720c5f8c2001629c9736608fac226c3957bc2408fc765a187ef7"} Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.785562 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17f9cd9539a4720c5f8c2001629c9736608fac226c3957bc2408fc765a187ef7" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.785630 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.809116 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b482j" event={"ID":"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc","Type":"ContainerStarted","Data":"e44db6133223dc93f4e83f53f1bf50ed6250d5f417bac0b1fbf2c6c5aec158e2"} Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.830120 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.830146 4823 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e3b2ffe1-5822-46ce-b1ee-23eb79596b97-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.830156 4823 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bad5782e-a906-462c-a585-c3fa56499827-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:08 crc kubenswrapper[4823]: I0320 09:01:08.830165 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bad5782e-a906-462c-a585-c3fa56499827-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:09 crc kubenswrapper[4823]: I0320 09:01:09.121611 4823 patch_prober.go:28] interesting pod/console-f9d7485db-lz4n9 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 20 09:01:09 crc kubenswrapper[4823]: I0320 09:01:09.121660 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-lz4n9" podUID="e104c875-7420-45b3-b675-a63b92c454e4" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 20 09:01:09 crc kubenswrapper[4823]: I0320 09:01:09.430570 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:01:09 crc kubenswrapper[4823]: I0320 09:01:09.430622 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:01:09 crc kubenswrapper[4823]: I0320 09:01:09.482457 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 09:01:09 crc kubenswrapper[4823]: I0320 09:01:09.499180 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-grjk6" Mar 20 09:01:15 crc kubenswrapper[4823]: I0320 09:01:15.636825 4823 ???:1] "http: TLS handshake error from 192.168.126.11:51052: no serving certificate available for the kubelet" Mar 20 09:01:16 crc kubenswrapper[4823]: E0320 09:01:16.396424 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 20 09:01:16 crc kubenswrapper[4823]: E0320 09:01:16.396653 4823 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 09:01:16 crc kubenswrapper[4823]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 20 09:01:16 crc kubenswrapper[4823]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ncldg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29566620-4lfkt_openshift-infra(65a02b02-7f51-4774-b050-292c3fd4f70b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 20 09:01:16 crc kubenswrapper[4823]: > logger="UnhandledError" Mar 20 09:01:16 crc kubenswrapper[4823]: E0320 09:01:16.398802 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29566620-4lfkt" podUID="65a02b02-7f51-4774-b050-292c3fd4f70b" Mar 20 09:01:16 crc kubenswrapper[4823]: E0320 09:01:16.853065 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29566620-4lfkt" podUID="65a02b02-7f51-4774-b050-292c3fd4f70b" Mar 20 09:01:17 crc kubenswrapper[4823]: I0320 09:01:17.079542 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-754fdbdd85-xt9c4"] Mar 20 09:01:17 crc kubenswrapper[4823]: I0320 09:01:17.079779 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" podUID="21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" containerName="controller-manager" containerID="cri-o://efd92f99adb9a9ac56756c38189daf7dab1234a10ca4b0ec1bad041e78d49b44" gracePeriod=30 Mar 20 09:01:17 crc kubenswrapper[4823]: I0320 09:01:17.096556 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj"] Mar 20 09:01:17 crc kubenswrapper[4823]: I0320 09:01:17.097090 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" podUID="fa5a375e-efce-4bf6-a949-37de5cd53861" containerName="route-controller-manager" containerID="cri-o://4a00ff2c23855f1d20e07bace0ab606b9b0a76f3906354f34917014e0a5665ea" gracePeriod=30 Mar 20 09:01:17 crc kubenswrapper[4823]: I0320 09:01:17.249145 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:01:17 crc kubenswrapper[4823]: I0320 09:01:17.859599 4823 generic.go:334] "Generic (PLEG): container finished" podID="fa5a375e-efce-4bf6-a949-37de5cd53861" containerID="4a00ff2c23855f1d20e07bace0ab606b9b0a76f3906354f34917014e0a5665ea" exitCode=0 Mar 20 09:01:17 crc kubenswrapper[4823]: I0320 09:01:17.859722 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" event={"ID":"fa5a375e-efce-4bf6-a949-37de5cd53861","Type":"ContainerDied","Data":"4a00ff2c23855f1d20e07bace0ab606b9b0a76f3906354f34917014e0a5665ea"} Mar 20 09:01:17 crc kubenswrapper[4823]: I0320 09:01:17.861647 4823 generic.go:334] "Generic (PLEG): container finished" podID="21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" containerID="efd92f99adb9a9ac56756c38189daf7dab1234a10ca4b0ec1bad041e78d49b44" exitCode=0 Mar 20 09:01:17 crc kubenswrapper[4823]: I0320 09:01:17.861692 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" event={"ID":"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6","Type":"ContainerDied","Data":"efd92f99adb9a9ac56756c38189daf7dab1234a10ca4b0ec1bad041e78d49b44"} Mar 20 09:01:18 crc kubenswrapper[4823]: I0320 09:01:18.486044 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:01:19 crc kubenswrapper[4823]: I0320 09:01:19.125384 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:01:19 crc kubenswrapper[4823]: I0320 09:01:19.128521 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:01:19 crc kubenswrapper[4823]: I0320 09:01:19.787298 4823 patch_prober.go:28] interesting pod/controller-manager-754fdbdd85-xt9c4 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": dial tcp 10.217.0.51:8443: connect: connection refused" start-of-body= Mar 20 09:01:19 crc kubenswrapper[4823]: I0320 09:01:19.787377 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" podUID="21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": dial tcp 10.217.0.51:8443: connect: connection refused" Mar 20 09:01:19 crc kubenswrapper[4823]: I0320 09:01:19.799774 4823 patch_prober.go:28] interesting pod/route-controller-manager-797bf5c8dc-clhqj container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.52:8443/healthz\": dial tcp 10.217.0.52:8443: connect: connection refused" start-of-body= Mar 20 09:01:19 crc kubenswrapper[4823]: I0320 09:01:19.800021 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" podUID="fa5a375e-efce-4bf6-a949-37de5cd53861" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.52:8443/healthz\": dial tcp 10.217.0.52:8443: connect: connection refused" Mar 20 09:01:24 crc kubenswrapper[4823]: E0320 09:01:24.860289 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 20 09:01:24 crc kubenswrapper[4823]: E0320 09:01:24.861021 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dtcp6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-k9bv8_openshift-marketplace(f9cc626b-cf2b-407f-ac25-c9ce01e0660d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:24 crc kubenswrapper[4823]: E0320 09:01:24.863059 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-k9bv8" podUID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" Mar 20 09:01:25 crc kubenswrapper[4823]: E0320 09:01:25.992832 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-k9bv8" podUID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.568657 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.569493 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l482s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-7g599_openshift-marketplace(95495063-7a29-4af1-a255-973a0129e289): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.570680 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-7g599" podUID="95495063-7a29-4af1-a255-973a0129e289" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.587670 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.593297 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.628994 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.629353 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c2df5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-sj5ms_openshift-marketplace(4fafad3e-2369-4997-beb3-c73d884b2276): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.630442 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q"] Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.630509 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-sj5ms" podUID="4fafad3e-2369-4997-beb3-c73d884b2276" Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.630705 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b2ffe1-5822-46ce-b1ee-23eb79596b97" containerName="pruner" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.630720 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b2ffe1-5822-46ce-b1ee-23eb79596b97" containerName="pruner" Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.630727 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad5782e-a906-462c-a585-c3fa56499827" containerName="pruner" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.630734 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad5782e-a906-462c-a585-c3fa56499827" containerName="pruner" Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.630772 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" containerName="controller-manager" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.630780 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" containerName="controller-manager" Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.630791 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa5a375e-efce-4bf6-a949-37de5cd53861" containerName="route-controller-manager" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.630797 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa5a375e-efce-4bf6-a949-37de5cd53861" containerName="route-controller-manager" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.630905 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad5782e-a906-462c-a585-c3fa56499827" containerName="pruner" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.630917 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa5a375e-efce-4bf6-a949-37de5cd53861" containerName="route-controller-manager" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.630927 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" containerName="controller-manager" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.630939 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b2ffe1-5822-46ce-b1ee-23eb79596b97" containerName="pruner" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.631321 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.645522 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q"] Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.655577 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.655753 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n5fpf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cbwgk_openshift-marketplace(a0ddc452-8cdd-4451-b2a3-215c24347335): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:27 crc kubenswrapper[4823]: E0320 09:01:27.657037 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cbwgk" podUID="a0ddc452-8cdd-4451-b2a3-215c24347335" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.692479 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-config\") pod \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.692558 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-client-ca\") pod \"fa5a375e-efce-4bf6-a949-37de5cd53861\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.692604 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-client-ca\") pod \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.692641 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-config\") pod \"fa5a375e-efce-4bf6-a949-37de5cd53861\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.692678 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-proxy-ca-bundles\") pod \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.692723 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-serving-cert\") pod \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.692787 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qptv8\" (UniqueName: \"kubernetes.io/projected/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-kube-api-access-qptv8\") pod \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\" (UID: \"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6\") " Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.692821 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwcnz\" (UniqueName: \"kubernetes.io/projected/fa5a375e-efce-4bf6-a949-37de5cd53861-kube-api-access-nwcnz\") pod \"fa5a375e-efce-4bf6-a949-37de5cd53861\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.692851 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa5a375e-efce-4bf6-a949-37de5cd53861-serving-cert\") pod \"fa5a375e-efce-4bf6-a949-37de5cd53861\" (UID: \"fa5a375e-efce-4bf6-a949-37de5cd53861\") " Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.694256 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-client-ca" (OuterVolumeSpecName: "client-ca") pod "fa5a375e-efce-4bf6-a949-37de5cd53861" (UID: "fa5a375e-efce-4bf6-a949-37de5cd53861"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.694651 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" (UID: "21d3c4ae-d6f7-421b-bf3d-f905a74b3db6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.694678 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-client-ca" (OuterVolumeSpecName: "client-ca") pod "21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" (UID: "21d3c4ae-d6f7-421b-bf3d-f905a74b3db6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.694953 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-config" (OuterVolumeSpecName: "config") pod "21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" (UID: "21d3c4ae-d6f7-421b-bf3d-f905a74b3db6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.695619 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-config" (OuterVolumeSpecName: "config") pod "fa5a375e-efce-4bf6-a949-37de5cd53861" (UID: "fa5a375e-efce-4bf6-a949-37de5cd53861"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.698926 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5a375e-efce-4bf6-a949-37de5cd53861-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fa5a375e-efce-4bf6-a949-37de5cd53861" (UID: "fa5a375e-efce-4bf6-a949-37de5cd53861"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.704957 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa5a375e-efce-4bf6-a949-37de5cd53861-kube-api-access-nwcnz" (OuterVolumeSpecName: "kube-api-access-nwcnz") pod "fa5a375e-efce-4bf6-a949-37de5cd53861" (UID: "fa5a375e-efce-4bf6-a949-37de5cd53861"). InnerVolumeSpecName "kube-api-access-nwcnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.705632 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-kube-api-access-qptv8" (OuterVolumeSpecName: "kube-api-access-qptv8") pod "21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" (UID: "21d3c4ae-d6f7-421b-bf3d-f905a74b3db6"). InnerVolumeSpecName "kube-api-access-qptv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.705834 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" (UID: "21d3c4ae-d6f7-421b-bf3d-f905a74b3db6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.793899 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-config\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.793946 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b2766f0-fe3e-430c-9def-6127425a735a-serving-cert\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.794002 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-client-ca\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.794036 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6sqf\" (UniqueName: \"kubernetes.io/projected/6b2766f0-fe3e-430c-9def-6127425a735a-kube-api-access-d6sqf\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.794102 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.794115 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.794123 4823 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.794131 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.794140 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qptv8\" (UniqueName: \"kubernetes.io/projected/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-kube-api-access-qptv8\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.794148 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwcnz\" (UniqueName: \"kubernetes.io/projected/fa5a375e-efce-4bf6-a949-37de5cd53861-kube-api-access-nwcnz\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.794156 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa5a375e-efce-4bf6-a949-37de5cd53861-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.794164 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.794171 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa5a375e-efce-4bf6-a949-37de5cd53861-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.896134 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-client-ca\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.897822 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-client-ca\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.898031 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6sqf\" (UniqueName: \"kubernetes.io/projected/6b2766f0-fe3e-430c-9def-6127425a735a-kube-api-access-d6sqf\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.898355 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-config\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.902904 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-config\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.898388 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b2766f0-fe3e-430c-9def-6127425a735a-serving-cert\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.911669 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b2766f0-fe3e-430c-9def-6127425a735a-serving-cert\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.923141 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" event={"ID":"21d3c4ae-d6f7-421b-bf3d-f905a74b3db6","Type":"ContainerDied","Data":"0dbc7949e2ef9cf323cfe05579325d13e82b32b0cd80544608b8e5ec3a4e7ce5"} Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.923230 4823 scope.go:117] "RemoveContainer" containerID="efd92f99adb9a9ac56756c38189daf7dab1234a10ca4b0ec1bad041e78d49b44" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.923368 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-754fdbdd85-xt9c4" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.925820 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6sqf\" (UniqueName: \"kubernetes.io/projected/6b2766f0-fe3e-430c-9def-6127425a735a-kube-api-access-d6sqf\") pod \"route-controller-manager-5bbbd4bd7-qlc5q\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.938192 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.939905 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj" event={"ID":"fa5a375e-efce-4bf6-a949-37de5cd53861","Type":"ContainerDied","Data":"f5c0f9108bb074c32d438269fbe7f0c07834317dba5659ddadaaa892e8f57be5"} Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.945512 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.991904 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj"] Mar 20 09:01:27 crc kubenswrapper[4823]: I0320 09:01:27.994781 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-797bf5c8dc-clhqj"] Mar 20 09:01:28 crc kubenswrapper[4823]: I0320 09:01:28.019091 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-754fdbdd85-xt9c4"] Mar 20 09:01:28 crc kubenswrapper[4823]: I0320 09:01:28.022262 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-754fdbdd85-xt9c4"] Mar 20 09:01:28 crc kubenswrapper[4823]: I0320 09:01:28.104987 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21d3c4ae-d6f7-421b-bf3d-f905a74b3db6" path="/var/lib/kubelet/pods/21d3c4ae-d6f7-421b-bf3d-f905a74b3db6/volumes" Mar 20 09:01:28 crc kubenswrapper[4823]: I0320 09:01:28.105869 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa5a375e-efce-4bf6-a949-37de5cd53861" path="/var/lib/kubelet/pods/fa5a375e-efce-4bf6-a949-37de5cd53861/volumes" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.421034 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5bcd788588-mqb2f"] Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.422572 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.427574 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.427894 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.428038 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.430558 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.430848 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.431008 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.432507 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5bcd788588-mqb2f"] Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.433680 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.545471 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9731aa2f-a26d-4575-8723-0509b8e467e9-serving-cert\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.545535 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsxq8\" (UniqueName: \"kubernetes.io/projected/9731aa2f-a26d-4575-8723-0509b8e467e9-kube-api-access-vsxq8\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.545663 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-config\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.545711 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-proxy-ca-bundles\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.545783 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-client-ca\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.647074 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsxq8\" (UniqueName: \"kubernetes.io/projected/9731aa2f-a26d-4575-8723-0509b8e467e9-kube-api-access-vsxq8\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.647154 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-config\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.647188 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-proxy-ca-bundles\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.647222 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-client-ca\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.647288 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9731aa2f-a26d-4575-8723-0509b8e467e9-serving-cert\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.648231 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-client-ca\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.648471 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-config\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.648532 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-proxy-ca-bundles\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.660600 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9731aa2f-a26d-4575-8723-0509b8e467e9-serving-cert\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.667580 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsxq8\" (UniqueName: \"kubernetes.io/projected/9731aa2f-a26d-4575-8723-0509b8e467e9-kube-api-access-vsxq8\") pod \"controller-manager-5bcd788588-mqb2f\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:30 crc kubenswrapper[4823]: I0320 09:01:30.744124 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.015719 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g2xtb" Mar 20 09:01:31 crc kubenswrapper[4823]: E0320 09:01:31.058525 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cbwgk" podUID="a0ddc452-8cdd-4451-b2a3-215c24347335" Mar 20 09:01:31 crc kubenswrapper[4823]: E0320 09:01:31.058579 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-sj5ms" podUID="4fafad3e-2369-4997-beb3-c73d884b2276" Mar 20 09:01:31 crc kubenswrapper[4823]: E0320 09:01:31.058644 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-7g599" podUID="95495063-7a29-4af1-a255-973a0129e289" Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.153322 4823 scope.go:117] "RemoveContainer" containerID="4a00ff2c23855f1d20e07bace0ab606b9b0a76f3906354f34917014e0a5665ea" Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.345607 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q"] Mar 20 09:01:31 crc kubenswrapper[4823]: W0320 09:01:31.374796 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b2766f0_fe3e_430c_9def_6127425a735a.slice/crio-772e04238d90d5ac67258651b9c7e450c0955e7d1a4232284d450c30b6de06c3 WatchSource:0}: Error finding container 772e04238d90d5ac67258651b9c7e450c0955e7d1a4232284d450c30b6de06c3: Status 404 returned error can't find the container with id 772e04238d90d5ac67258651b9c7e450c0955e7d1a4232284d450c30b6de06c3 Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.464689 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5bcd788588-mqb2f"] Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.958597 4823 generic.go:334] "Generic (PLEG): container finished" podID="2cee23f5-1268-4adf-8466-da5931de9825" containerID="af7f3dcbbfcda86f18af6a5e208c415b74b95b55440662f61f73fe293e2cd466" exitCode=0 Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.958941 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m5xl" event={"ID":"2cee23f5-1268-4adf-8466-da5931de9825","Type":"ContainerDied","Data":"af7f3dcbbfcda86f18af6a5e208c415b74b95b55440662f61f73fe293e2cd466"} Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.961462 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" event={"ID":"9731aa2f-a26d-4575-8723-0509b8e467e9","Type":"ContainerStarted","Data":"b8ea55280c84fa8332dbc6cb0d47b84b12c87ed48160209877c5117b05fe1a86"} Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.961518 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" event={"ID":"9731aa2f-a26d-4575-8723-0509b8e467e9","Type":"ContainerStarted","Data":"19a8ed1155bb9b3955c1f64a8e4ded55e50a838235e2f37c1ce8f55471c1724f"} Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.961768 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.963969 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566620-4lfkt" event={"ID":"65a02b02-7f51-4774-b050-292c3fd4f70b","Type":"ContainerStarted","Data":"417b495fce674061bd7abe4a9f7535e74794235255d9303d98800d9ec849cc3c"} Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.965840 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.967446 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" event={"ID":"6b2766f0-fe3e-430c-9def-6127425a735a","Type":"ContainerStarted","Data":"215e7c04be562fbc2b38cca9aaacbb3a77f6359619a2f59c1bcc0ef7de7b90e8"} Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.967496 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" event={"ID":"6b2766f0-fe3e-430c-9def-6127425a735a","Type":"ContainerStarted","Data":"772e04238d90d5ac67258651b9c7e450c0955e7d1a4232284d450c30b6de06c3"} Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.967673 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.969540 4823 generic.go:334] "Generic (PLEG): container finished" podID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" containerID="3a09c00eb4b76dd622ade74f197e954040923bb85a0f89d2de253c928dcbed53" exitCode=0 Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.969603 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-425tz" event={"ID":"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe","Type":"ContainerDied","Data":"3a09c00eb4b76dd622ade74f197e954040923bb85a0f89d2de253c928dcbed53"} Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.971774 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbgxd" event={"ID":"694805b6-f8a1-4277-8132-100503fdc1a8","Type":"ContainerStarted","Data":"a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5"} Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.972115 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.973189 4823 generic.go:334] "Generic (PLEG): container finished" podID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerID="3e9bf8df7e91a5a85aae24c8f69a000b48e79c446ee7ca20de5d9077f78e44d9" exitCode=0 Mar 20 09:01:31 crc kubenswrapper[4823]: I0320 09:01:31.973214 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b482j" event={"ID":"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc","Type":"ContainerDied","Data":"3e9bf8df7e91a5a85aae24c8f69a000b48e79c446ee7ca20de5d9077f78e44d9"} Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.006820 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" podStartSLOduration=15.006800054 podStartE2EDuration="15.006800054s" podCreationTimestamp="2026-03-20 09:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:32.006144896 +0000 UTC m=+258.692041280" watchObservedRunningTime="2026-03-20 09:01:32.006800054 +0000 UTC m=+258.692696428" Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.086699 4823 csr.go:261] certificate signing request csr-dctfz is approved, waiting to be issued Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.087639 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" podStartSLOduration=15.08762357 podStartE2EDuration="15.08762357s" podCreationTimestamp="2026-03-20 09:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:32.086979872 +0000 UTC m=+258.772876246" watchObservedRunningTime="2026-03-20 09:01:32.08762357 +0000 UTC m=+258.773519934" Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.094492 4823 csr.go:257] certificate signing request csr-dctfz is issued Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.130941 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566620-4lfkt" podStartSLOduration=54.239884144 podStartE2EDuration="1m32.130923733s" podCreationTimestamp="2026-03-20 09:00:00 +0000 UTC" firstStartedPulling="2026-03-20 09:00:53.373089855 +0000 UTC m=+220.058986239" lastFinishedPulling="2026-03-20 09:01:31.264129444 +0000 UTC m=+257.950025828" observedRunningTime="2026-03-20 09:01:32.126655826 +0000 UTC m=+258.812552200" watchObservedRunningTime="2026-03-20 09:01:32.130923733 +0000 UTC m=+258.816820107" Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.694765 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.695585 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.697687 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.697759 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.697827 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.873993 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.874099 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.975301 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.975453 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.975553 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.980355 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-425tz" event={"ID":"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe","Type":"ContainerStarted","Data":"216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca"} Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.984956 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m5xl" event={"ID":"2cee23f5-1268-4adf-8466-da5931de9825","Type":"ContainerStarted","Data":"9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a"} Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.989838 4823 generic.go:334] "Generic (PLEG): container finished" podID="65a02b02-7f51-4774-b050-292c3fd4f70b" containerID="417b495fce674061bd7abe4a9f7535e74794235255d9303d98800d9ec849cc3c" exitCode=0 Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.989900 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566620-4lfkt" event={"ID":"65a02b02-7f51-4774-b050-292c3fd4f70b","Type":"ContainerDied","Data":"417b495fce674061bd7abe4a9f7535e74794235255d9303d98800d9ec849cc3c"} Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.992833 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b482j" event={"ID":"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc","Type":"ContainerStarted","Data":"96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a"} Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.995973 4823 generic.go:334] "Generic (PLEG): container finished" podID="694805b6-f8a1-4277-8132-100503fdc1a8" containerID="a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5" exitCode=0 Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.997187 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbgxd" event={"ID":"694805b6-f8a1-4277-8132-100503fdc1a8","Type":"ContainerDied","Data":"a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5"} Mar 20 09:01:32 crc kubenswrapper[4823]: I0320 09:01:32.998408 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:33 crc kubenswrapper[4823]: I0320 09:01:33.009861 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-425tz" podStartSLOduration=3.179678459 podStartE2EDuration="34.009842126s" podCreationTimestamp="2026-03-20 09:00:59 +0000 UTC" firstStartedPulling="2026-03-20 09:01:01.601227524 +0000 UTC m=+228.287123898" lastFinishedPulling="2026-03-20 09:01:32.431391181 +0000 UTC m=+259.117287565" observedRunningTime="2026-03-20 09:01:33.007181943 +0000 UTC m=+259.693078317" watchObservedRunningTime="2026-03-20 09:01:33.009842126 +0000 UTC m=+259.695738500" Mar 20 09:01:33 crc kubenswrapper[4823]: I0320 09:01:33.041437 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:33 crc kubenswrapper[4823]: I0320 09:01:33.063116 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6m5xl" podStartSLOduration=10.107904896 podStartE2EDuration="34.063061262s" podCreationTimestamp="2026-03-20 09:00:59 +0000 UTC" firstStartedPulling="2026-03-20 09:01:08.620712325 +0000 UTC m=+235.306608699" lastFinishedPulling="2026-03-20 09:01:32.575868651 +0000 UTC m=+259.261765065" observedRunningTime="2026-03-20 09:01:33.058978291 +0000 UTC m=+259.744874665" watchObservedRunningTime="2026-03-20 09:01:33.063061262 +0000 UTC m=+259.748957636" Mar 20 09:01:33 crc kubenswrapper[4823]: I0320 09:01:33.098650 4823 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-18 16:05:59.852730042 +0000 UTC Mar 20 09:01:33 crc kubenswrapper[4823]: I0320 09:01:33.098696 4823 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7303h4m26.754037497s for next certificate rotation Mar 20 09:01:33 crc kubenswrapper[4823]: I0320 09:01:33.456045 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 09:01:34 crc kubenswrapper[4823]: I0320 09:01:34.001553 4823 generic.go:334] "Generic (PLEG): container finished" podID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerID="96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a" exitCode=0 Mar 20 09:01:34 crc kubenswrapper[4823]: I0320 09:01:34.001649 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b482j" event={"ID":"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc","Type":"ContainerDied","Data":"96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a"} Mar 20 09:01:34 crc kubenswrapper[4823]: I0320 09:01:34.004732 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad","Type":"ContainerStarted","Data":"faa4f2d4cd220ea4018919eff72610df2dff390bcf4e222f6f803ff94f45768e"} Mar 20 09:01:34 crc kubenswrapper[4823]: I0320 09:01:34.004780 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad","Type":"ContainerStarted","Data":"4daf90b6d94433b1ef685faa7ab5cf07140568188b02272ecc856ae38d29cc49"} Mar 20 09:01:34 crc kubenswrapper[4823]: I0320 09:01:34.045316 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.045295493 podStartE2EDuration="2.045295493s" podCreationTimestamp="2026-03-20 09:01:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:34.040284754 +0000 UTC m=+260.726181128" watchObservedRunningTime="2026-03-20 09:01:34.045295493 +0000 UTC m=+260.731191877" Mar 20 09:01:34 crc kubenswrapper[4823]: I0320 09:01:34.099399 4823 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-20 20:55:13.638185651 +0000 UTC Mar 20 09:01:34 crc kubenswrapper[4823]: I0320 09:01:34.099440 4823 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6611h53m39.538749167s for next certificate rotation Mar 20 09:01:34 crc kubenswrapper[4823]: I0320 09:01:34.305914 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566620-4lfkt" Mar 20 09:01:34 crc kubenswrapper[4823]: I0320 09:01:34.394706 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncldg\" (UniqueName: \"kubernetes.io/projected/65a02b02-7f51-4774-b050-292c3fd4f70b-kube-api-access-ncldg\") pod \"65a02b02-7f51-4774-b050-292c3fd4f70b\" (UID: \"65a02b02-7f51-4774-b050-292c3fd4f70b\") " Mar 20 09:01:34 crc kubenswrapper[4823]: I0320 09:01:34.408089 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65a02b02-7f51-4774-b050-292c3fd4f70b-kube-api-access-ncldg" (OuterVolumeSpecName: "kube-api-access-ncldg") pod "65a02b02-7f51-4774-b050-292c3fd4f70b" (UID: "65a02b02-7f51-4774-b050-292c3fd4f70b"). InnerVolumeSpecName "kube-api-access-ncldg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:34 crc kubenswrapper[4823]: I0320 09:01:34.495935 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncldg\" (UniqueName: \"kubernetes.io/projected/65a02b02-7f51-4774-b050-292c3fd4f70b-kube-api-access-ncldg\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:35 crc kubenswrapper[4823]: I0320 09:01:35.012543 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbgxd" event={"ID":"694805b6-f8a1-4277-8132-100503fdc1a8","Type":"ContainerStarted","Data":"27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569"} Mar 20 09:01:35 crc kubenswrapper[4823]: I0320 09:01:35.013900 4823 generic.go:334] "Generic (PLEG): container finished" podID="9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad" containerID="faa4f2d4cd220ea4018919eff72610df2dff390bcf4e222f6f803ff94f45768e" exitCode=0 Mar 20 09:01:35 crc kubenswrapper[4823]: I0320 09:01:35.013936 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad","Type":"ContainerDied","Data":"faa4f2d4cd220ea4018919eff72610df2dff390bcf4e222f6f803ff94f45768e"} Mar 20 09:01:35 crc kubenswrapper[4823]: I0320 09:01:35.015308 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566620-4lfkt" event={"ID":"65a02b02-7f51-4774-b050-292c3fd4f70b","Type":"ContainerDied","Data":"0dc23f2f6bb8904a31284e767413949b1f157d46ea17e175f7e2e9db9585cc69"} Mar 20 09:01:35 crc kubenswrapper[4823]: I0320 09:01:35.015357 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0dc23f2f6bb8904a31284e767413949b1f157d46ea17e175f7e2e9db9585cc69" Mar 20 09:01:35 crc kubenswrapper[4823]: I0320 09:01:35.015372 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566620-4lfkt" Mar 20 09:01:35 crc kubenswrapper[4823]: I0320 09:01:35.064523 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vbgxd" podStartSLOduration=9.659323328 podStartE2EDuration="35.064508422s" podCreationTimestamp="2026-03-20 09:01:00 +0000 UTC" firstStartedPulling="2026-03-20 09:01:08.620705984 +0000 UTC m=+235.306602359" lastFinishedPulling="2026-03-20 09:01:34.025891079 +0000 UTC m=+260.711787453" observedRunningTime="2026-03-20 09:01:35.043901434 +0000 UTC m=+261.729797828" watchObservedRunningTime="2026-03-20 09:01:35.064508422 +0000 UTC m=+261.750404786" Mar 20 09:01:36 crc kubenswrapper[4823]: I0320 09:01:36.277964 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:36 crc kubenswrapper[4823]: I0320 09:01:36.418986 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kubelet-dir\") pod \"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad\" (UID: \"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad\") " Mar 20 09:01:36 crc kubenswrapper[4823]: I0320 09:01:36.419058 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kube-api-access\") pod \"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad\" (UID: \"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad\") " Mar 20 09:01:36 crc kubenswrapper[4823]: I0320 09:01:36.419132 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad" (UID: "9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:01:36 crc kubenswrapper[4823]: I0320 09:01:36.419531 4823 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:36 crc kubenswrapper[4823]: I0320 09:01:36.424262 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad" (UID: "9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:36 crc kubenswrapper[4823]: I0320 09:01:36.520579 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:37 crc kubenswrapper[4823]: I0320 09:01:37.028215 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad","Type":"ContainerDied","Data":"4daf90b6d94433b1ef685faa7ab5cf07140568188b02272ecc856ae38d29cc49"} Mar 20 09:01:37 crc kubenswrapper[4823]: I0320 09:01:37.028278 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4daf90b6d94433b1ef685faa7ab5cf07140568188b02272ecc856ae38d29cc49" Mar 20 09:01:37 crc kubenswrapper[4823]: I0320 09:01:37.028361 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:37 crc kubenswrapper[4823]: I0320 09:01:37.097803 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5bcd788588-mqb2f"] Mar 20 09:01:37 crc kubenswrapper[4823]: I0320 09:01:37.098058 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" podUID="9731aa2f-a26d-4575-8723-0509b8e467e9" containerName="controller-manager" containerID="cri-o://b8ea55280c84fa8332dbc6cb0d47b84b12c87ed48160209877c5117b05fe1a86" gracePeriod=30 Mar 20 09:01:37 crc kubenswrapper[4823]: I0320 09:01:37.193667 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q"] Mar 20 09:01:37 crc kubenswrapper[4823]: I0320 09:01:37.194242 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" podUID="6b2766f0-fe3e-430c-9def-6127425a735a" containerName="route-controller-manager" containerID="cri-o://215e7c04be562fbc2b38cca9aaacbb3a77f6359619a2f59c1bcc0ef7de7b90e8" gracePeriod=30 Mar 20 09:01:37 crc kubenswrapper[4823]: I0320 09:01:37.946819 4823 patch_prober.go:28] interesting pod/route-controller-manager-5bbbd4bd7-qlc5q container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Mar 20 09:01:37 crc kubenswrapper[4823]: I0320 09:01:37.946868 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" podUID="6b2766f0-fe3e-430c-9def-6127425a735a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.036416 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b482j" event={"ID":"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc","Type":"ContainerStarted","Data":"a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c"} Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.038294 4823 generic.go:334] "Generic (PLEG): container finished" podID="6b2766f0-fe3e-430c-9def-6127425a735a" containerID="215e7c04be562fbc2b38cca9aaacbb3a77f6359619a2f59c1bcc0ef7de7b90e8" exitCode=0 Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.038373 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" event={"ID":"6b2766f0-fe3e-430c-9def-6127425a735a","Type":"ContainerDied","Data":"215e7c04be562fbc2b38cca9aaacbb3a77f6359619a2f59c1bcc0ef7de7b90e8"} Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.040801 4823 generic.go:334] "Generic (PLEG): container finished" podID="9731aa2f-a26d-4575-8723-0509b8e467e9" containerID="b8ea55280c84fa8332dbc6cb0d47b84b12c87ed48160209877c5117b05fe1a86" exitCode=0 Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.040834 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" event={"ID":"9731aa2f-a26d-4575-8723-0509b8e467e9","Type":"ContainerDied","Data":"b8ea55280c84fa8332dbc6cb0d47b84b12c87ed48160209877c5117b05fe1a86"} Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.064643 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b482j" podStartSLOduration=31.68409318 podStartE2EDuration="37.064606869s" podCreationTimestamp="2026-03-20 09:01:01 +0000 UTC" firstStartedPulling="2026-03-20 09:01:31.975697247 +0000 UTC m=+258.661593621" lastFinishedPulling="2026-03-20 09:01:37.356210936 +0000 UTC m=+264.042107310" observedRunningTime="2026-03-20 09:01:38.061156327 +0000 UTC m=+264.747052701" watchObservedRunningTime="2026-03-20 09:01:38.064606869 +0000 UTC m=+264.750503243" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.328161 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.356063 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6cdcd76459-72ph6"] Mar 20 09:01:38 crc kubenswrapper[4823]: E0320 09:01:38.356267 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9731aa2f-a26d-4575-8723-0509b8e467e9" containerName="controller-manager" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.356280 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="9731aa2f-a26d-4575-8723-0509b8e467e9" containerName="controller-manager" Mar 20 09:01:38 crc kubenswrapper[4823]: E0320 09:01:38.356292 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65a02b02-7f51-4774-b050-292c3fd4f70b" containerName="oc" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.356299 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="65a02b02-7f51-4774-b050-292c3fd4f70b" containerName="oc" Mar 20 09:01:38 crc kubenswrapper[4823]: E0320 09:01:38.356310 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad" containerName="pruner" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.356317 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad" containerName="pruner" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.356430 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="9731aa2f-a26d-4575-8723-0509b8e467e9" containerName="controller-manager" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.356445 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="65a02b02-7f51-4774-b050-292c3fd4f70b" containerName="oc" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.356458 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8c76b9-a5ca-42f0-b847-ecd71de7b9ad" containerName="pruner" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.356926 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.372361 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6cdcd76459-72ph6"] Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.445268 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-client-ca\") pod \"9731aa2f-a26d-4575-8723-0509b8e467e9\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.445512 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-config\") pod \"9731aa2f-a26d-4575-8723-0509b8e467e9\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.445578 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-proxy-ca-bundles\") pod \"9731aa2f-a26d-4575-8723-0509b8e467e9\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.445610 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsxq8\" (UniqueName: \"kubernetes.io/projected/9731aa2f-a26d-4575-8723-0509b8e467e9-kube-api-access-vsxq8\") pod \"9731aa2f-a26d-4575-8723-0509b8e467e9\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.445686 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9731aa2f-a26d-4575-8723-0509b8e467e9-serving-cert\") pod \"9731aa2f-a26d-4575-8723-0509b8e467e9\" (UID: \"9731aa2f-a26d-4575-8723-0509b8e467e9\") " Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.445975 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-config\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.446068 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-client-ca\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.446099 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af274430-1b44-4cf7-836a-96786249ac28-serving-cert\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.446169 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn42c\" (UniqueName: \"kubernetes.io/projected/af274430-1b44-4cf7-836a-96786249ac28-kube-api-access-vn42c\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.446272 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9731aa2f-a26d-4575-8723-0509b8e467e9" (UID: "9731aa2f-a26d-4575-8723-0509b8e467e9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.446337 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-proxy-ca-bundles\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.446403 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-config" (OuterVolumeSpecName: "config") pod "9731aa2f-a26d-4575-8723-0509b8e467e9" (UID: "9731aa2f-a26d-4575-8723-0509b8e467e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.446538 4823 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.446796 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-client-ca" (OuterVolumeSpecName: "client-ca") pod "9731aa2f-a26d-4575-8723-0509b8e467e9" (UID: "9731aa2f-a26d-4575-8723-0509b8e467e9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.450785 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9731aa2f-a26d-4575-8723-0509b8e467e9-kube-api-access-vsxq8" (OuterVolumeSpecName: "kube-api-access-vsxq8") pod "9731aa2f-a26d-4575-8723-0509b8e467e9" (UID: "9731aa2f-a26d-4575-8723-0509b8e467e9"). InnerVolumeSpecName "kube-api-access-vsxq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.462459 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9731aa2f-a26d-4575-8723-0509b8e467e9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9731aa2f-a26d-4575-8723-0509b8e467e9" (UID: "9731aa2f-a26d-4575-8723-0509b8e467e9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.548006 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-client-ca\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.548063 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af274430-1b44-4cf7-836a-96786249ac28-serving-cert\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.548118 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn42c\" (UniqueName: \"kubernetes.io/projected/af274430-1b44-4cf7-836a-96786249ac28-kube-api-access-vn42c\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.548156 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-proxy-ca-bundles\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.548227 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-config\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.548306 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.548322 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsxq8\" (UniqueName: \"kubernetes.io/projected/9731aa2f-a26d-4575-8723-0509b8e467e9-kube-api-access-vsxq8\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.548337 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9731aa2f-a26d-4575-8723-0509b8e467e9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.548348 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9731aa2f-a26d-4575-8723-0509b8e467e9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.549925 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-config\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.550659 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-client-ca\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.552406 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-proxy-ca-bundles\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.559652 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af274430-1b44-4cf7-836a-96786249ac28-serving-cert\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.612003 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn42c\" (UniqueName: \"kubernetes.io/projected/af274430-1b44-4cf7-836a-96786249ac28-kube-api-access-vn42c\") pod \"controller-manager-6cdcd76459-72ph6\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:38 crc kubenswrapper[4823]: I0320 09:01:38.670965 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.041498 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.048064 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" event={"ID":"9731aa2f-a26d-4575-8723-0509b8e467e9","Type":"ContainerDied","Data":"19a8ed1155bb9b3955c1f64a8e4ded55e50a838235e2f37c1ce8f55471c1724f"} Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.048145 4823 scope.go:117] "RemoveContainer" containerID="b8ea55280c84fa8332dbc6cb0d47b84b12c87ed48160209877c5117b05fe1a86" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.048411 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bcd788588-mqb2f" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.051378 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.051505 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q" event={"ID":"6b2766f0-fe3e-430c-9def-6127425a735a","Type":"ContainerDied","Data":"772e04238d90d5ac67258651b9c7e450c0955e7d1a4232284d450c30b6de06c3"} Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.066948 4823 scope.go:117] "RemoveContainer" containerID="215e7c04be562fbc2b38cca9aaacbb3a77f6359619a2f59c1bcc0ef7de7b90e8" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.098491 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5bcd788588-mqb2f"] Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.105187 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5bcd788588-mqb2f"] Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.129250 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6cdcd76459-72ph6"] Mar 20 09:01:39 crc kubenswrapper[4823]: W0320 09:01:39.142097 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf274430_1b44_4cf7_836a_96786249ac28.slice/crio-1cc7c639dda76be1bd4a23cbcef6f07badb6657790b31bc1a0ddf50f4cf6c459 WatchSource:0}: Error finding container 1cc7c639dda76be1bd4a23cbcef6f07badb6657790b31bc1a0ddf50f4cf6c459: Status 404 returned error can't find the container with id 1cc7c639dda76be1bd4a23cbcef6f07badb6657790b31bc1a0ddf50f4cf6c459 Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.153912 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-config\") pod \"6b2766f0-fe3e-430c-9def-6127425a735a\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.154036 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-client-ca\") pod \"6b2766f0-fe3e-430c-9def-6127425a735a\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.154128 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b2766f0-fe3e-430c-9def-6127425a735a-serving-cert\") pod \"6b2766f0-fe3e-430c-9def-6127425a735a\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.154218 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6sqf\" (UniqueName: \"kubernetes.io/projected/6b2766f0-fe3e-430c-9def-6127425a735a-kube-api-access-d6sqf\") pod \"6b2766f0-fe3e-430c-9def-6127425a735a\" (UID: \"6b2766f0-fe3e-430c-9def-6127425a735a\") " Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.155272 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-client-ca" (OuterVolumeSpecName: "client-ca") pod "6b2766f0-fe3e-430c-9def-6127425a735a" (UID: "6b2766f0-fe3e-430c-9def-6127425a735a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.155354 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-config" (OuterVolumeSpecName: "config") pod "6b2766f0-fe3e-430c-9def-6127425a735a" (UID: "6b2766f0-fe3e-430c-9def-6127425a735a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.159452 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b2766f0-fe3e-430c-9def-6127425a735a-kube-api-access-d6sqf" (OuterVolumeSpecName: "kube-api-access-d6sqf") pod "6b2766f0-fe3e-430c-9def-6127425a735a" (UID: "6b2766f0-fe3e-430c-9def-6127425a735a"). InnerVolumeSpecName "kube-api-access-d6sqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.160004 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2766f0-fe3e-430c-9def-6127425a735a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6b2766f0-fe3e-430c-9def-6127425a735a" (UID: "6b2766f0-fe3e-430c-9def-6127425a735a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.256502 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.256824 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b2766f0-fe3e-430c-9def-6127425a735a-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.256838 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b2766f0-fe3e-430c-9def-6127425a735a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.256852 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6sqf\" (UniqueName: \"kubernetes.io/projected/6b2766f0-fe3e-430c-9def-6127425a735a-kube-api-access-d6sqf\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.298724 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 09:01:39 crc kubenswrapper[4823]: E0320 09:01:39.299054 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2766f0-fe3e-430c-9def-6127425a735a" containerName="route-controller-manager" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.299077 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2766f0-fe3e-430c-9def-6127425a735a" containerName="route-controller-manager" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.299213 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b2766f0-fe3e-430c-9def-6127425a735a" containerName="route-controller-manager" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.299701 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.302152 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.309021 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.309069 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.379015 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q"] Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.390269 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd4bd7-qlc5q"] Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.431507 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.431574 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.458206 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c826ac6f-8206-4ce3-a59a-c03afefbc038-kube-api-access\") pod \"installer-9-crc\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.458286 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.458308 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-var-lock\") pod \"installer-9-crc\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.559332 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c826ac6f-8206-4ce3-a59a-c03afefbc038-kube-api-access\") pod \"installer-9-crc\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.559400 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.559421 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-var-lock\") pod \"installer-9-crc\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.559489 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-var-lock\") pod \"installer-9-crc\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.559839 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.580840 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c826ac6f-8206-4ce3-a59a-c03afefbc038-kube-api-access\") pod \"installer-9-crc\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.649899 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.819795 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:01:39 crc kubenswrapper[4823]: I0320 09:01:39.820197 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.064531 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" event={"ID":"af274430-1b44-4cf7-836a-96786249ac28","Type":"ContainerStarted","Data":"cbf50b2f670b73197747697eaac24eea240de04640472306a125d6ce9d71c66b"} Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.064610 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" event={"ID":"af274430-1b44-4cf7-836a-96786249ac28","Type":"ContainerStarted","Data":"1cc7c639dda76be1bd4a23cbcef6f07badb6657790b31bc1a0ddf50f4cf6c459"} Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.064778 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.068690 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 09:01:40 crc kubenswrapper[4823]: W0320 09:01:40.083940 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc826ac6f_8206_4ce3_a59a_c03afefbc038.slice/crio-1fbe640b66ab8c698502de38b9d8a1e61cf48c9e6ef66730b223fa8b0d05509b WatchSource:0}: Error finding container 1fbe640b66ab8c698502de38b9d8a1e61cf48c9e6ef66730b223fa8b0d05509b: Status 404 returned error can't find the container with id 1fbe640b66ab8c698502de38b9d8a1e61cf48c9e6ef66730b223fa8b0d05509b Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.101256 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" podStartSLOduration=3.101240875 podStartE2EDuration="3.101240875s" podCreationTimestamp="2026-03-20 09:01:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:40.099171544 +0000 UTC m=+266.785067918" watchObservedRunningTime="2026-03-20 09:01:40.101240875 +0000 UTC m=+266.787137249" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.109727 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b2766f0-fe3e-430c-9def-6127425a735a" path="/var/lib/kubelet/pods/6b2766f0-fe3e-430c-9def-6127425a735a/volumes" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.110868 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9731aa2f-a26d-4575-8723-0509b8e467e9" path="/var/lib/kubelet/pods/9731aa2f-a26d-4575-8723-0509b8e467e9/volumes" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.111528 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.211938 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.212500 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.424577 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.426835 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.431077 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz"] Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.431873 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.443630 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.443728 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.456179 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.456404 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.463988 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.464029 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.462901 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz"] Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.518997 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.575401 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-client-ca\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.575569 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbk8r\" (UniqueName: \"kubernetes.io/projected/ba099115-fe14-4477-b503-bfc6a4d078c9-kube-api-access-gbk8r\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.575639 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-config\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.575795 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba099115-fe14-4477-b503-bfc6a4d078c9-serving-cert\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.677226 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbk8r\" (UniqueName: \"kubernetes.io/projected/ba099115-fe14-4477-b503-bfc6a4d078c9-kube-api-access-gbk8r\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.677286 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-config\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.677363 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba099115-fe14-4477-b503-bfc6a4d078c9-serving-cert\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.677393 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-client-ca\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.678365 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-client-ca\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.678623 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-config\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.682821 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba099115-fe14-4477-b503-bfc6a4d078c9-serving-cert\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.696644 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbk8r\" (UniqueName: \"kubernetes.io/projected/ba099115-fe14-4477-b503-bfc6a4d078c9-kube-api-access-gbk8r\") pod \"route-controller-manager-59c8fb7fc-zphbz\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.749686 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:40 crc kubenswrapper[4823]: I0320 09:01:40.980070 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz"] Mar 20 09:01:41 crc kubenswrapper[4823]: I0320 09:01:41.059704 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:41 crc kubenswrapper[4823]: I0320 09:01:41.059781 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:41 crc kubenswrapper[4823]: I0320 09:01:41.093782 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" event={"ID":"ba099115-fe14-4477-b503-bfc6a4d078c9","Type":"ContainerStarted","Data":"9439b120d7a5472d646813eeb6d2bdd498b4046b4ce3346213041ad8c5e4ce92"} Mar 20 09:01:41 crc kubenswrapper[4823]: I0320 09:01:41.108228 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9bv8" event={"ID":"f9cc626b-cf2b-407f-ac25-c9ce01e0660d","Type":"ContainerStarted","Data":"e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc"} Mar 20 09:01:41 crc kubenswrapper[4823]: I0320 09:01:41.110687 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c826ac6f-8206-4ce3-a59a-c03afefbc038","Type":"ContainerStarted","Data":"a9885ecfdcf3152d8ac207eea93d988198d761502409927563c405c09edbd829"} Mar 20 09:01:41 crc kubenswrapper[4823]: I0320 09:01:41.110777 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c826ac6f-8206-4ce3-a59a-c03afefbc038","Type":"ContainerStarted","Data":"1fbe640b66ab8c698502de38b9d8a1e61cf48c9e6ef66730b223fa8b0d05509b"} Mar 20 09:01:41 crc kubenswrapper[4823]: I0320 09:01:41.137938 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.137914365 podStartE2EDuration="2.137914365s" podCreationTimestamp="2026-03-20 09:01:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:41.132037061 +0000 UTC m=+267.817933425" watchObservedRunningTime="2026-03-20 09:01:41.137914365 +0000 UTC m=+267.823810739" Mar 20 09:01:41 crc kubenswrapper[4823]: I0320 09:01:41.154518 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:41 crc kubenswrapper[4823]: I0320 09:01:41.478103 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:41 crc kubenswrapper[4823]: I0320 09:01:41.478537 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:42 crc kubenswrapper[4823]: I0320 09:01:42.083327 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-425tz"] Mar 20 09:01:42 crc kubenswrapper[4823]: I0320 09:01:42.105564 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vbgxd" podUID="694805b6-f8a1-4277-8132-100503fdc1a8" containerName="registry-server" probeResult="failure" output=< Mar 20 09:01:42 crc kubenswrapper[4823]: timeout: failed to connect service ":50051" within 1s Mar 20 09:01:42 crc kubenswrapper[4823]: > Mar 20 09:01:42 crc kubenswrapper[4823]: I0320 09:01:42.118299 4823 generic.go:334] "Generic (PLEG): container finished" podID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" containerID="e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc" exitCode=0 Mar 20 09:01:42 crc kubenswrapper[4823]: I0320 09:01:42.118426 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9bv8" event={"ID":"f9cc626b-cf2b-407f-ac25-c9ce01e0660d","Type":"ContainerDied","Data":"e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc"} Mar 20 09:01:42 crc kubenswrapper[4823]: I0320 09:01:42.119894 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" event={"ID":"ba099115-fe14-4477-b503-bfc6a4d078c9","Type":"ContainerStarted","Data":"803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998"} Mar 20 09:01:42 crc kubenswrapper[4823]: I0320 09:01:42.120323 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:42 crc kubenswrapper[4823]: I0320 09:01:42.129312 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:42 crc kubenswrapper[4823]: I0320 09:01:42.163716 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" podStartSLOduration=5.163694764 podStartE2EDuration="5.163694764s" podCreationTimestamp="2026-03-20 09:01:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:42.163083595 +0000 UTC m=+268.848979969" watchObservedRunningTime="2026-03-20 09:01:42.163694764 +0000 UTC m=+268.849591138" Mar 20 09:01:42 crc kubenswrapper[4823]: I0320 09:01:42.521549 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b482j" podUID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerName="registry-server" probeResult="failure" output=< Mar 20 09:01:42 crc kubenswrapper[4823]: timeout: failed to connect service ":50051" within 1s Mar 20 09:01:42 crc kubenswrapper[4823]: > Mar 20 09:01:43 crc kubenswrapper[4823]: I0320 09:01:43.130285 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9bv8" event={"ID":"f9cc626b-cf2b-407f-ac25-c9ce01e0660d","Type":"ContainerStarted","Data":"0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00"} Mar 20 09:01:43 crc kubenswrapper[4823]: I0320 09:01:43.130434 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-425tz" podUID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" containerName="registry-server" containerID="cri-o://216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca" gracePeriod=2 Mar 20 09:01:43 crc kubenswrapper[4823]: I0320 09:01:43.163957 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k9bv8" podStartSLOduration=2.9544978950000003 podStartE2EDuration="46.163940257s" podCreationTimestamp="2026-03-20 09:00:57 +0000 UTC" firstStartedPulling="2026-03-20 09:00:59.326724763 +0000 UTC m=+226.012621137" lastFinishedPulling="2026-03-20 09:01:42.536167115 +0000 UTC m=+269.222063499" observedRunningTime="2026-03-20 09:01:43.16165316 +0000 UTC m=+269.847549524" watchObservedRunningTime="2026-03-20 09:01:43.163940257 +0000 UTC m=+269.849836631" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.136183 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.136996 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbwgk" event={"ID":"a0ddc452-8cdd-4451-b2a3-215c24347335","Type":"ContainerStarted","Data":"de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001"} Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.139178 4823 generic.go:334] "Generic (PLEG): container finished" podID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" containerID="216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca" exitCode=0 Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.139235 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-425tz" event={"ID":"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe","Type":"ContainerDied","Data":"216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca"} Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.139259 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-425tz" event={"ID":"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe","Type":"ContainerDied","Data":"229a0693bfd2db0add9998f62ebf35d8efbfc8c149524af77853ae2ca42c766a"} Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.139278 4823 scope.go:117] "RemoveContainer" containerID="216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.139396 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-425tz" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.146248 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj5ms" event={"ID":"4fafad3e-2369-4997-beb3-c73d884b2276","Type":"ContainerStarted","Data":"5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9"} Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.160893 4823 scope.go:117] "RemoveContainer" containerID="3a09c00eb4b76dd622ade74f197e954040923bb85a0f89d2de253c928dcbed53" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.192424 4823 scope.go:117] "RemoveContainer" containerID="c2f21152a51a4ef4355aaa4144fd73da9cf09e78184cf4f9e558105251a18be4" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.229064 4823 scope.go:117] "RemoveContainer" containerID="216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca" Mar 20 09:01:44 crc kubenswrapper[4823]: E0320 09:01:44.230987 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca\": container with ID starting with 216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca not found: ID does not exist" containerID="216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.231037 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca"} err="failed to get container status \"216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca\": rpc error: code = NotFound desc = could not find container \"216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca\": container with ID starting with 216de77dc7ca3d501d631828d3ffcb3efc49c42fa03bc6d8e7d806d1ec95acca not found: ID does not exist" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.231064 4823 scope.go:117] "RemoveContainer" containerID="3a09c00eb4b76dd622ade74f197e954040923bb85a0f89d2de253c928dcbed53" Mar 20 09:01:44 crc kubenswrapper[4823]: E0320 09:01:44.231406 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a09c00eb4b76dd622ade74f197e954040923bb85a0f89d2de253c928dcbed53\": container with ID starting with 3a09c00eb4b76dd622ade74f197e954040923bb85a0f89d2de253c928dcbed53 not found: ID does not exist" containerID="3a09c00eb4b76dd622ade74f197e954040923bb85a0f89d2de253c928dcbed53" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.231445 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a09c00eb4b76dd622ade74f197e954040923bb85a0f89d2de253c928dcbed53"} err="failed to get container status \"3a09c00eb4b76dd622ade74f197e954040923bb85a0f89d2de253c928dcbed53\": rpc error: code = NotFound desc = could not find container \"3a09c00eb4b76dd622ade74f197e954040923bb85a0f89d2de253c928dcbed53\": container with ID starting with 3a09c00eb4b76dd622ade74f197e954040923bb85a0f89d2de253c928dcbed53 not found: ID does not exist" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.231463 4823 scope.go:117] "RemoveContainer" containerID="c2f21152a51a4ef4355aaa4144fd73da9cf09e78184cf4f9e558105251a18be4" Mar 20 09:01:44 crc kubenswrapper[4823]: E0320 09:01:44.231832 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2f21152a51a4ef4355aaa4144fd73da9cf09e78184cf4f9e558105251a18be4\": container with ID starting with c2f21152a51a4ef4355aaa4144fd73da9cf09e78184cf4f9e558105251a18be4 not found: ID does not exist" containerID="c2f21152a51a4ef4355aaa4144fd73da9cf09e78184cf4f9e558105251a18be4" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.231920 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f21152a51a4ef4355aaa4144fd73da9cf09e78184cf4f9e558105251a18be4"} err="failed to get container status \"c2f21152a51a4ef4355aaa4144fd73da9cf09e78184cf4f9e558105251a18be4\": rpc error: code = NotFound desc = could not find container \"c2f21152a51a4ef4355aaa4144fd73da9cf09e78184cf4f9e558105251a18be4\": container with ID starting with c2f21152a51a4ef4355aaa4144fd73da9cf09e78184cf4f9e558105251a18be4 not found: ID does not exist" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.236049 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7772\" (UniqueName: \"kubernetes.io/projected/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-kube-api-access-t7772\") pod \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.236104 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-catalog-content\") pod \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.236240 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-utilities\") pod \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\" (UID: \"86c6eee3-c57f-41af-8bf5-73f1c15ca2fe\") " Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.241952 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-utilities" (OuterVolumeSpecName: "utilities") pod "86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" (UID: "86c6eee3-c57f-41af-8bf5-73f1c15ca2fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.265593 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-kube-api-access-t7772" (OuterVolumeSpecName: "kube-api-access-t7772") pod "86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" (UID: "86c6eee3-c57f-41af-8bf5-73f1c15ca2fe"). InnerVolumeSpecName "kube-api-access-t7772". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.270222 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" (UID: "86c6eee3-c57f-41af-8bf5-73f1c15ca2fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.338162 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.338204 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7772\" (UniqueName: \"kubernetes.io/projected/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-kube-api-access-t7772\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.338218 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.466016 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-425tz"] Mar 20 09:01:44 crc kubenswrapper[4823]: I0320 09:01:44.468848 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-425tz"] Mar 20 09:01:45 crc kubenswrapper[4823]: I0320 09:01:45.154993 4823 generic.go:334] "Generic (PLEG): container finished" podID="a0ddc452-8cdd-4451-b2a3-215c24347335" containerID="de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001" exitCode=0 Mar 20 09:01:45 crc kubenswrapper[4823]: I0320 09:01:45.155044 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbwgk" event={"ID":"a0ddc452-8cdd-4451-b2a3-215c24347335","Type":"ContainerDied","Data":"de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001"} Mar 20 09:01:45 crc kubenswrapper[4823]: I0320 09:01:45.171778 4823 generic.go:334] "Generic (PLEG): container finished" podID="4fafad3e-2369-4997-beb3-c73d884b2276" containerID="5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9" exitCode=0 Mar 20 09:01:45 crc kubenswrapper[4823]: I0320 09:01:45.171852 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj5ms" event={"ID":"4fafad3e-2369-4997-beb3-c73d884b2276","Type":"ContainerDied","Data":"5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9"} Mar 20 09:01:46 crc kubenswrapper[4823]: I0320 09:01:46.106064 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" path="/var/lib/kubelet/pods/86c6eee3-c57f-41af-8bf5-73f1c15ca2fe/volumes" Mar 20 09:01:46 crc kubenswrapper[4823]: I0320 09:01:46.178874 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj5ms" event={"ID":"4fafad3e-2369-4997-beb3-c73d884b2276","Type":"ContainerStarted","Data":"98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf"} Mar 20 09:01:47 crc kubenswrapper[4823]: I0320 09:01:47.204872 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sj5ms" podStartSLOduration=4.018955425 podStartE2EDuration="50.204853393s" podCreationTimestamp="2026-03-20 09:00:57 +0000 UTC" firstStartedPulling="2026-03-20 09:00:59.397695748 +0000 UTC m=+226.083592122" lastFinishedPulling="2026-03-20 09:01:45.583593716 +0000 UTC m=+272.269490090" observedRunningTime="2026-03-20 09:01:47.203232435 +0000 UTC m=+273.889128809" watchObservedRunningTime="2026-03-20 09:01:47.204853393 +0000 UTC m=+273.890749767" Mar 20 09:01:48 crc kubenswrapper[4823]: I0320 09:01:48.190674 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbwgk" event={"ID":"a0ddc452-8cdd-4451-b2a3-215c24347335","Type":"ContainerStarted","Data":"72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c"} Mar 20 09:01:48 crc kubenswrapper[4823]: I0320 09:01:48.215971 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:01:48 crc kubenswrapper[4823]: I0320 09:01:48.216029 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:01:48 crc kubenswrapper[4823]: I0320 09:01:48.224523 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:01:48 crc kubenswrapper[4823]: I0320 09:01:48.225825 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:01:48 crc kubenswrapper[4823]: I0320 09:01:48.269709 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:01:48 crc kubenswrapper[4823]: I0320 09:01:48.287840 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:01:49 crc kubenswrapper[4823]: I0320 09:01:49.226784 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cbwgk" podStartSLOduration=3.627679804 podStartE2EDuration="51.226761484s" podCreationTimestamp="2026-03-20 09:00:58 +0000 UTC" firstStartedPulling="2026-03-20 09:00:59.342708083 +0000 UTC m=+226.028604457" lastFinishedPulling="2026-03-20 09:01:46.941789753 +0000 UTC m=+273.627686137" observedRunningTime="2026-03-20 09:01:49.222529048 +0000 UTC m=+275.908425462" watchObservedRunningTime="2026-03-20 09:01:49.226761484 +0000 UTC m=+275.912657878" Mar 20 09:01:49 crc kubenswrapper[4823]: I0320 09:01:49.243765 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:01:50 crc kubenswrapper[4823]: I0320 09:01:50.201635 4823 generic.go:334] "Generic (PLEG): container finished" podID="95495063-7a29-4af1-a255-973a0129e289" containerID="cc290a0fed4218f661268d816343115152c92147297093a88527cf98238659eb" exitCode=0 Mar 20 09:01:50 crc kubenswrapper[4823]: I0320 09:01:50.201709 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g599" event={"ID":"95495063-7a29-4af1-a255-973a0129e289","Type":"ContainerDied","Data":"cc290a0fed4218f661268d816343115152c92147297093a88527cf98238659eb"} Mar 20 09:01:51 crc kubenswrapper[4823]: I0320 09:01:51.129919 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:51 crc kubenswrapper[4823]: I0320 09:01:51.205521 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:01:51 crc kubenswrapper[4823]: I0320 09:01:51.525099 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:51 crc kubenswrapper[4823]: I0320 09:01:51.568037 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:52 crc kubenswrapper[4823]: I0320 09:01:52.220163 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g599" event={"ID":"95495063-7a29-4af1-a255-973a0129e289","Type":"ContainerStarted","Data":"f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf"} Mar 20 09:01:52 crc kubenswrapper[4823]: I0320 09:01:52.254238 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7g599" podStartSLOduration=3.568500299 podStartE2EDuration="55.254205004s" podCreationTimestamp="2026-03-20 09:00:57 +0000 UTC" firstStartedPulling="2026-03-20 09:00:59.453269119 +0000 UTC m=+226.139165493" lastFinishedPulling="2026-03-20 09:01:51.138973814 +0000 UTC m=+277.824870198" observedRunningTime="2026-03-20 09:01:52.249451804 +0000 UTC m=+278.935348198" watchObservedRunningTime="2026-03-20 09:01:52.254205004 +0000 UTC m=+278.940101418" Mar 20 09:01:54 crc kubenswrapper[4823]: I0320 09:01:54.686443 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b482j"] Mar 20 09:01:54 crc kubenswrapper[4823]: I0320 09:01:54.687182 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b482j" podUID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerName="registry-server" containerID="cri-o://a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c" gracePeriod=2 Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.196660 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.242013 4823 generic.go:334] "Generic (PLEG): container finished" podID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerID="a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c" exitCode=0 Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.242067 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b482j" event={"ID":"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc","Type":"ContainerDied","Data":"a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c"} Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.242109 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b482j" event={"ID":"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc","Type":"ContainerDied","Data":"e44db6133223dc93f4e83f53f1bf50ed6250d5f417bac0b1fbf2c6c5aec158e2"} Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.242139 4823 scope.go:117] "RemoveContainer" containerID="a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.242291 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b482j" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.275885 4823 scope.go:117] "RemoveContainer" containerID="96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.289096 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pswd\" (UniqueName: \"kubernetes.io/projected/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-kube-api-access-4pswd\") pod \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.289259 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-catalog-content\") pod \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.289299 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-utilities\") pod \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\" (UID: \"de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc\") " Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.291066 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-utilities" (OuterVolumeSpecName: "utilities") pod "de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" (UID: "de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.298285 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-kube-api-access-4pswd" (OuterVolumeSpecName: "kube-api-access-4pswd") pod "de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" (UID: "de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc"). InnerVolumeSpecName "kube-api-access-4pswd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.305094 4823 scope.go:117] "RemoveContainer" containerID="3e9bf8df7e91a5a85aae24c8f69a000b48e79c446ee7ca20de5d9077f78e44d9" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.344299 4823 scope.go:117] "RemoveContainer" containerID="a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c" Mar 20 09:01:55 crc kubenswrapper[4823]: E0320 09:01:55.344959 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c\": container with ID starting with a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c not found: ID does not exist" containerID="a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.345174 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c"} err="failed to get container status \"a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c\": rpc error: code = NotFound desc = could not find container \"a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c\": container with ID starting with a1e9201da2d5c0785caea69dba9d5f9eb4704a95afca95e5ae3bb7d22724cd1c not found: ID does not exist" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.345337 4823 scope.go:117] "RemoveContainer" containerID="96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a" Mar 20 09:01:55 crc kubenswrapper[4823]: E0320 09:01:55.346311 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a\": container with ID starting with 96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a not found: ID does not exist" containerID="96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.346361 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a"} err="failed to get container status \"96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a\": rpc error: code = NotFound desc = could not find container \"96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a\": container with ID starting with 96ece196c68c8d27edb1b51495c1baeca056a8c1248be8ffdf902990bcba156a not found: ID does not exist" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.346383 4823 scope.go:117] "RemoveContainer" containerID="3e9bf8df7e91a5a85aae24c8f69a000b48e79c446ee7ca20de5d9077f78e44d9" Mar 20 09:01:55 crc kubenswrapper[4823]: E0320 09:01:55.346822 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e9bf8df7e91a5a85aae24c8f69a000b48e79c446ee7ca20de5d9077f78e44d9\": container with ID starting with 3e9bf8df7e91a5a85aae24c8f69a000b48e79c446ee7ca20de5d9077f78e44d9 not found: ID does not exist" containerID="3e9bf8df7e91a5a85aae24c8f69a000b48e79c446ee7ca20de5d9077f78e44d9" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.347130 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e9bf8df7e91a5a85aae24c8f69a000b48e79c446ee7ca20de5d9077f78e44d9"} err="failed to get container status \"3e9bf8df7e91a5a85aae24c8f69a000b48e79c446ee7ca20de5d9077f78e44d9\": rpc error: code = NotFound desc = could not find container \"3e9bf8df7e91a5a85aae24c8f69a000b48e79c446ee7ca20de5d9077f78e44d9\": container with ID starting with 3e9bf8df7e91a5a85aae24c8f69a000b48e79c446ee7ca20de5d9077f78e44d9 not found: ID does not exist" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.393454 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pswd\" (UniqueName: \"kubernetes.io/projected/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-kube-api-access-4pswd\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.393973 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.457343 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" (UID: "de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.495634 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.599011 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b482j"] Mar 20 09:01:55 crc kubenswrapper[4823]: I0320 09:01:55.605308 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b482j"] Mar 20 09:01:56 crc kubenswrapper[4823]: I0320 09:01:56.104514 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" path="/var/lib/kubelet/pods/de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc/volumes" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.091889 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6cdcd76459-72ph6"] Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.092126 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" podUID="af274430-1b44-4cf7-836a-96786249ac28" containerName="controller-manager" containerID="cri-o://cbf50b2f670b73197747697eaac24eea240de04640472306a125d6ce9d71c66b" gracePeriod=30 Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.135892 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz"] Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.138316 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" podUID="ba099115-fe14-4477-b503-bfc6a4d078c9" containerName="route-controller-manager" containerID="cri-o://803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998" gracePeriod=30 Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.254600 4823 generic.go:334] "Generic (PLEG): container finished" podID="af274430-1b44-4cf7-836a-96786249ac28" containerID="cbf50b2f670b73197747697eaac24eea240de04640472306a125d6ce9d71c66b" exitCode=0 Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.254922 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" event={"ID":"af274430-1b44-4cf7-836a-96786249ac28","Type":"ContainerDied","Data":"cbf50b2f670b73197747697eaac24eea240de04640472306a125d6ce9d71c66b"} Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.628755 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.679469 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.726266 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-config\") pod \"ba099115-fe14-4477-b503-bfc6a4d078c9\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.726401 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-client-ca\") pod \"ba099115-fe14-4477-b503-bfc6a4d078c9\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.726548 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba099115-fe14-4477-b503-bfc6a4d078c9-serving-cert\") pod \"ba099115-fe14-4477-b503-bfc6a4d078c9\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.726849 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbk8r\" (UniqueName: \"kubernetes.io/projected/ba099115-fe14-4477-b503-bfc6a4d078c9-kube-api-access-gbk8r\") pod \"ba099115-fe14-4477-b503-bfc6a4d078c9\" (UID: \"ba099115-fe14-4477-b503-bfc6a4d078c9\") " Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.727491 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-config" (OuterVolumeSpecName: "config") pod "ba099115-fe14-4477-b503-bfc6a4d078c9" (UID: "ba099115-fe14-4477-b503-bfc6a4d078c9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.727512 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-client-ca" (OuterVolumeSpecName: "client-ca") pod "ba099115-fe14-4477-b503-bfc6a4d078c9" (UID: "ba099115-fe14-4477-b503-bfc6a4d078c9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.732476 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba099115-fe14-4477-b503-bfc6a4d078c9-kube-api-access-gbk8r" (OuterVolumeSpecName: "kube-api-access-gbk8r") pod "ba099115-fe14-4477-b503-bfc6a4d078c9" (UID: "ba099115-fe14-4477-b503-bfc6a4d078c9"). InnerVolumeSpecName "kube-api-access-gbk8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.732955 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba099115-fe14-4477-b503-bfc6a4d078c9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ba099115-fe14-4477-b503-bfc6a4d078c9" (UID: "ba099115-fe14-4477-b503-bfc6a4d078c9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.816020 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.816498 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.828684 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-client-ca\") pod \"af274430-1b44-4cf7-836a-96786249ac28\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.828810 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn42c\" (UniqueName: \"kubernetes.io/projected/af274430-1b44-4cf7-836a-96786249ac28-kube-api-access-vn42c\") pod \"af274430-1b44-4cf7-836a-96786249ac28\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.828883 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-proxy-ca-bundles\") pod \"af274430-1b44-4cf7-836a-96786249ac28\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.828925 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-config\") pod \"af274430-1b44-4cf7-836a-96786249ac28\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.829694 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "af274430-1b44-4cf7-836a-96786249ac28" (UID: "af274430-1b44-4cf7-836a-96786249ac28"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.829021 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af274430-1b44-4cf7-836a-96786249ac28-serving-cert\") pod \"af274430-1b44-4cf7-836a-96786249ac28\" (UID: \"af274430-1b44-4cf7-836a-96786249ac28\") " Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.829959 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-config" (OuterVolumeSpecName: "config") pod "af274430-1b44-4cf7-836a-96786249ac28" (UID: "af274430-1b44-4cf7-836a-96786249ac28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.830211 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-client-ca" (OuterVolumeSpecName: "client-ca") pod "af274430-1b44-4cf7-836a-96786249ac28" (UID: "af274430-1b44-4cf7-836a-96786249ac28"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.830392 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.830426 4823 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.830447 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba099115-fe14-4477-b503-bfc6a4d078c9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.830468 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.830488 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba099115-fe14-4477-b503-bfc6a4d078c9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.830508 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbk8r\" (UniqueName: \"kubernetes.io/projected/ba099115-fe14-4477-b503-bfc6a4d078c9-kube-api-access-gbk8r\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.830526 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af274430-1b44-4cf7-836a-96786249ac28-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.832830 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af274430-1b44-4cf7-836a-96786249ac28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "af274430-1b44-4cf7-836a-96786249ac28" (UID: "af274430-1b44-4cf7-836a-96786249ac28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.833003 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af274430-1b44-4cf7-836a-96786249ac28-kube-api-access-vn42c" (OuterVolumeSpecName: "kube-api-access-vn42c") pod "af274430-1b44-4cf7-836a-96786249ac28" (UID: "af274430-1b44-4cf7-836a-96786249ac28"). InnerVolumeSpecName "kube-api-access-vn42c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.877132 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.932991 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af274430-1b44-4cf7-836a-96786249ac28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4823]: I0320 09:01:57.933044 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn42c\" (UniqueName: \"kubernetes.io/projected/af274430-1b44-4cf7-836a-96786249ac28-kube-api-access-vn42c\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.264553 4823 generic.go:334] "Generic (PLEG): container finished" podID="ba099115-fe14-4477-b503-bfc6a4d078c9" containerID="803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998" exitCode=0 Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.264630 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" event={"ID":"ba099115-fe14-4477-b503-bfc6a4d078c9","Type":"ContainerDied","Data":"803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998"} Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.264660 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" event={"ID":"ba099115-fe14-4477-b503-bfc6a4d078c9","Type":"ContainerDied","Data":"9439b120d7a5472d646813eeb6d2bdd498b4046b4ce3346213041ad8c5e4ce92"} Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.264680 4823 scope.go:117] "RemoveContainer" containerID="803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.264805 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.271951 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.272024 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdcd76459-72ph6" event={"ID":"af274430-1b44-4cf7-836a-96786249ac28","Type":"ContainerDied","Data":"1cc7c639dda76be1bd4a23cbcef6f07badb6657790b31bc1a0ddf50f4cf6c459"} Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.278898 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.318511 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6cdcd76459-72ph6"] Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.321344 4823 scope.go:117] "RemoveContainer" containerID="803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.322195 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6cdcd76459-72ph6"] Mar 20 09:01:58 crc kubenswrapper[4823]: E0320 09:01:58.322242 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998\": container with ID starting with 803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998 not found: ID does not exist" containerID="803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.322268 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998"} err="failed to get container status \"803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998\": rpc error: code = NotFound desc = could not find container \"803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998\": container with ID starting with 803cfdf4f0956a1bdaf9a78ac5493ab4b64d3f179dc7cbbf913d5783d29c1998 not found: ID does not exist" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.322288 4823 scope.go:117] "RemoveContainer" containerID="cbf50b2f670b73197747697eaac24eea240de04640472306a125d6ce9d71c66b" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.329789 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz"] Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.333266 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59c8fb7fc-zphbz"] Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.335435 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.400360 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.400894 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.448973 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2"] Mar 20 09:01:58 crc kubenswrapper[4823]: E0320 09:01:58.449831 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af274430-1b44-4cf7-836a-96786249ac28" containerName="controller-manager" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.450065 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="af274430-1b44-4cf7-836a-96786249ac28" containerName="controller-manager" Mar 20 09:01:58 crc kubenswrapper[4823]: E0320 09:01:58.450348 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" containerName="extract-content" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.450533 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" containerName="extract-content" Mar 20 09:01:58 crc kubenswrapper[4823]: E0320 09:01:58.450711 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerName="extract-utilities" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.450960 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerName="extract-utilities" Mar 20 09:01:58 crc kubenswrapper[4823]: E0320 09:01:58.451154 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerName="extract-content" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.451311 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerName="extract-content" Mar 20 09:01:58 crc kubenswrapper[4823]: E0320 09:01:58.451473 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" containerName="registry-server" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.451626 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" containerName="registry-server" Mar 20 09:01:58 crc kubenswrapper[4823]: E0320 09:01:58.451869 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" containerName="extract-utilities" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.452039 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" containerName="extract-utilities" Mar 20 09:01:58 crc kubenswrapper[4823]: E0320 09:01:58.452205 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerName="registry-server" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.452415 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerName="registry-server" Mar 20 09:01:58 crc kubenswrapper[4823]: E0320 09:01:58.452581 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba099115-fe14-4477-b503-bfc6a4d078c9" containerName="route-controller-manager" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.452837 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba099115-fe14-4477-b503-bfc6a4d078c9" containerName="route-controller-manager" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.454774 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="86c6eee3-c57f-41af-8bf5-73f1c15ca2fe" containerName="registry-server" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.455046 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="af274430-1b44-4cf7-836a-96786249ac28" containerName="controller-manager" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.455479 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba099115-fe14-4477-b503-bfc6a4d078c9" containerName="route-controller-manager" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.455832 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9e8a09-5d12-42e9-aba9-7fb5ec3cc6fc" containerName="registry-server" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.456943 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7"] Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.457074 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.462604 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.462730 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.462920 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.463107 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.463120 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.463805 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.463828 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7"] Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.463896 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.464415 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.465469 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2"] Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.465813 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.465858 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.466346 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.466438 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.466556 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.467093 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.471369 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.544235 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddd258bb-2c2e-4211-94b3-947051b719f0-serving-cert\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.544311 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-client-ca\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.544337 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-proxy-ca-bundles\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.544412 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-config\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.544448 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b77pd\" (UniqueName: \"kubernetes.io/projected/ddd258bb-2c2e-4211-94b3-947051b719f0-kube-api-access-b77pd\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.645874 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-client-ca\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.645969 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-config\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.646020 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-proxy-ca-bundles\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.646181 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-config\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.646227 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b77pd\" (UniqueName: \"kubernetes.io/projected/ddd258bb-2c2e-4211-94b3-947051b719f0-kube-api-access-b77pd\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.646264 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-client-ca\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.646289 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddd258bb-2c2e-4211-94b3-947051b719f0-serving-cert\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.646310 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-serving-cert\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.646380 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd7pd\" (UniqueName: \"kubernetes.io/projected/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-kube-api-access-jd7pd\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.646614 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-client-ca\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.647570 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-proxy-ca-bundles\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.647885 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-config\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.651591 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddd258bb-2c2e-4211-94b3-947051b719f0-serving-cert\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.663122 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b77pd\" (UniqueName: \"kubernetes.io/projected/ddd258bb-2c2e-4211-94b3-947051b719f0-kube-api-access-b77pd\") pod \"controller-manager-6bffb6f9f6-4sqz2\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.681321 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sj5ms"] Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.748002 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-client-ca\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.748057 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-serving-cert\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.748094 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd7pd\" (UniqueName: \"kubernetes.io/projected/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-kube-api-access-jd7pd\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.748119 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-config\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.749309 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-config\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.749418 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-client-ca\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.754510 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-serving-cert\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.780214 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd7pd\" (UniqueName: \"kubernetes.io/projected/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-kube-api-access-jd7pd\") pod \"route-controller-manager-5c54d7bcb-p4xv7\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.784605 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:01:58 crc kubenswrapper[4823]: I0320 09:01:58.794694 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.228785 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2"] Mar 20 09:01:59 crc kubenswrapper[4823]: W0320 09:01:59.241242 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddd258bb_2c2e_4211_94b3_947051b719f0.slice/crio-7d93edb4ac6fde0ecf3e0b5b477b5d314633253a6877fab6c30955eb4a38846c WatchSource:0}: Error finding container 7d93edb4ac6fde0ecf3e0b5b477b5d314633253a6877fab6c30955eb4a38846c: Status 404 returned error can't find the container with id 7d93edb4ac6fde0ecf3e0b5b477b5d314633253a6877fab6c30955eb4a38846c Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.257568 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7"] Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.289194 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" event={"ID":"ddd258bb-2c2e-4211-94b3-947051b719f0","Type":"ContainerStarted","Data":"7d93edb4ac6fde0ecf3e0b5b477b5d314633253a6877fab6c30955eb4a38846c"} Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.291194 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sj5ms" podUID="4fafad3e-2369-4997-beb3-c73d884b2276" containerName="registry-server" containerID="cri-o://98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf" gracePeriod=2 Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.335024 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.713462 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.870831 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-catalog-content\") pod \"4fafad3e-2369-4997-beb3-c73d884b2276\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.870930 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-utilities\") pod \"4fafad3e-2369-4997-beb3-c73d884b2276\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.870967 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2df5\" (UniqueName: \"kubernetes.io/projected/4fafad3e-2369-4997-beb3-c73d884b2276-kube-api-access-c2df5\") pod \"4fafad3e-2369-4997-beb3-c73d884b2276\" (UID: \"4fafad3e-2369-4997-beb3-c73d884b2276\") " Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.872569 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-utilities" (OuterVolumeSpecName: "utilities") pod "4fafad3e-2369-4997-beb3-c73d884b2276" (UID: "4fafad3e-2369-4997-beb3-c73d884b2276"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.876042 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fafad3e-2369-4997-beb3-c73d884b2276-kube-api-access-c2df5" (OuterVolumeSpecName: "kube-api-access-c2df5") pod "4fafad3e-2369-4997-beb3-c73d884b2276" (UID: "4fafad3e-2369-4997-beb3-c73d884b2276"). InnerVolumeSpecName "kube-api-access-c2df5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.921831 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4fafad3e-2369-4997-beb3-c73d884b2276" (UID: "4fafad3e-2369-4997-beb3-c73d884b2276"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.972648 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.972690 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fafad3e-2369-4997-beb3-c73d884b2276-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:59 crc kubenswrapper[4823]: I0320 09:01:59.972700 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2df5\" (UniqueName: \"kubernetes.io/projected/4fafad3e-2369-4997-beb3-c73d884b2276-kube-api-access-c2df5\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.104259 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af274430-1b44-4cf7-836a-96786249ac28" path="/var/lib/kubelet/pods/af274430-1b44-4cf7-836a-96786249ac28/volumes" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.105025 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba099115-fe14-4477-b503-bfc6a4d078c9" path="/var/lib/kubelet/pods/ba099115-fe14-4477-b503-bfc6a4d078c9/volumes" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.130035 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566622-nc9qz"] Mar 20 09:02:00 crc kubenswrapper[4823]: E0320 09:02:00.130301 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fafad3e-2369-4997-beb3-c73d884b2276" containerName="registry-server" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.130318 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fafad3e-2369-4997-beb3-c73d884b2276" containerName="registry-server" Mar 20 09:02:00 crc kubenswrapper[4823]: E0320 09:02:00.130335 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fafad3e-2369-4997-beb3-c73d884b2276" containerName="extract-utilities" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.130343 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fafad3e-2369-4997-beb3-c73d884b2276" containerName="extract-utilities" Mar 20 09:02:00 crc kubenswrapper[4823]: E0320 09:02:00.130356 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fafad3e-2369-4997-beb3-c73d884b2276" containerName="extract-content" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.130364 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fafad3e-2369-4997-beb3-c73d884b2276" containerName="extract-content" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.130469 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fafad3e-2369-4997-beb3-c73d884b2276" containerName="registry-server" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.130898 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566622-nc9qz" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.132808 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.133684 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.133824 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.141820 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566622-nc9qz"] Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.275973 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x448\" (UniqueName: \"kubernetes.io/projected/56782166-f1af-439a-b19b-8eb59caaffe9-kube-api-access-8x448\") pod \"auto-csr-approver-29566622-nc9qz\" (UID: \"56782166-f1af-439a-b19b-8eb59caaffe9\") " pod="openshift-infra/auto-csr-approver-29566622-nc9qz" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.297943 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" event={"ID":"2fca0d9b-55e0-49ae-a768-b2d1c19f1635","Type":"ContainerStarted","Data":"99ed05a7e22440487990806107359d6a57e851f13d011a823103b233707c4569"} Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.298003 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" event={"ID":"2fca0d9b-55e0-49ae-a768-b2d1c19f1635","Type":"ContainerStarted","Data":"fff6ac43622a18efe15cfddc42e0fa727a34af6d1c90ba28888593eb67d2d7f6"} Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.298110 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.301592 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" event={"ID":"ddd258bb-2c2e-4211-94b3-947051b719f0","Type":"ContainerStarted","Data":"8fa7b75a76a8e3f3c0dc367d0bf6dc097b3a9526fe7cd13738a217d3b18882ad"} Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.301802 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.305388 4823 generic.go:334] "Generic (PLEG): container finished" podID="4fafad3e-2369-4997-beb3-c73d884b2276" containerID="98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf" exitCode=0 Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.305475 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj5ms" event={"ID":"4fafad3e-2369-4997-beb3-c73d884b2276","Type":"ContainerDied","Data":"98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf"} Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.305499 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sj5ms" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.305527 4823 scope.go:117] "RemoveContainer" containerID="98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.305511 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sj5ms" event={"ID":"4fafad3e-2369-4997-beb3-c73d884b2276","Type":"ContainerDied","Data":"2d02cbcbf3b3f8cad7228ae4c76ce24d4928bd90d48d49b900d0e907150dab46"} Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.307262 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.307957 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.323008 4823 scope.go:117] "RemoveContainer" containerID="5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.323017 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" podStartSLOduration=3.32299117 podStartE2EDuration="3.32299117s" podCreationTimestamp="2026-03-20 09:01:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:02:00.316060306 +0000 UTC m=+287.001956720" watchObservedRunningTime="2026-03-20 09:02:00.32299117 +0000 UTC m=+287.008887564" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.341504 4823 scope.go:117] "RemoveContainer" containerID="47dc9a523f536c8f14d995f072b38e6c61ee7c4bd45af89cd627bab79f7bd90a" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.354055 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sj5ms"] Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.367886 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sj5ms"] Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.372433 4823 scope.go:117] "RemoveContainer" containerID="98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf" Mar 20 09:02:00 crc kubenswrapper[4823]: E0320 09:02:00.373083 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf\": container with ID starting with 98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf not found: ID does not exist" containerID="98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.373112 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf"} err="failed to get container status \"98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf\": rpc error: code = NotFound desc = could not find container \"98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf\": container with ID starting with 98dc024c392cf57fb9e364cf123d878e9b76db1c4ac311d6d5934be28ebf53cf not found: ID does not exist" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.373135 4823 scope.go:117] "RemoveContainer" containerID="5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9" Mar 20 09:02:00 crc kubenswrapper[4823]: E0320 09:02:00.373693 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9\": container with ID starting with 5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9 not found: ID does not exist" containerID="5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.373735 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9"} err="failed to get container status \"5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9\": rpc error: code = NotFound desc = could not find container \"5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9\": container with ID starting with 5d2d77f0929c569802eccc76faae218c4f99b3bf69a30c68cdacb74fac9dbca9 not found: ID does not exist" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.373774 4823 scope.go:117] "RemoveContainer" containerID="47dc9a523f536c8f14d995f072b38e6c61ee7c4bd45af89cd627bab79f7bd90a" Mar 20 09:02:00 crc kubenswrapper[4823]: E0320 09:02:00.374394 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47dc9a523f536c8f14d995f072b38e6c61ee7c4bd45af89cd627bab79f7bd90a\": container with ID starting with 47dc9a523f536c8f14d995f072b38e6c61ee7c4bd45af89cd627bab79f7bd90a not found: ID does not exist" containerID="47dc9a523f536c8f14d995f072b38e6c61ee7c4bd45af89cd627bab79f7bd90a" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.374435 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47dc9a523f536c8f14d995f072b38e6c61ee7c4bd45af89cd627bab79f7bd90a"} err="failed to get container status \"47dc9a523f536c8f14d995f072b38e6c61ee7c4bd45af89cd627bab79f7bd90a\": rpc error: code = NotFound desc = could not find container \"47dc9a523f536c8f14d995f072b38e6c61ee7c4bd45af89cd627bab79f7bd90a\": container with ID starting with 47dc9a523f536c8f14d995f072b38e6c61ee7c4bd45af89cd627bab79f7bd90a not found: ID does not exist" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.377123 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x448\" (UniqueName: \"kubernetes.io/projected/56782166-f1af-439a-b19b-8eb59caaffe9-kube-api-access-8x448\") pod \"auto-csr-approver-29566622-nc9qz\" (UID: \"56782166-f1af-439a-b19b-8eb59caaffe9\") " pod="openshift-infra/auto-csr-approver-29566622-nc9qz" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.382025 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" podStartSLOduration=3.3820122440000002 podStartE2EDuration="3.382012244s" podCreationTimestamp="2026-03-20 09:01:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:02:00.3805112 +0000 UTC m=+287.066407584" watchObservedRunningTime="2026-03-20 09:02:00.382012244 +0000 UTC m=+287.067908618" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.399542 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x448\" (UniqueName: \"kubernetes.io/projected/56782166-f1af-439a-b19b-8eb59caaffe9-kube-api-access-8x448\") pod \"auto-csr-approver-29566622-nc9qz\" (UID: \"56782166-f1af-439a-b19b-8eb59caaffe9\") " pod="openshift-infra/auto-csr-approver-29566622-nc9qz" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.448013 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566622-nc9qz" Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.628248 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566622-nc9qz"] Mar 20 09:02:00 crc kubenswrapper[4823]: W0320 09:02:00.636152 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56782166_f1af_439a_b19b_8eb59caaffe9.slice/crio-45c0170dc5ad836b310ff6aeaa4d4826fdffac5ffd500b23147ebd6b3e8d2486 WatchSource:0}: Error finding container 45c0170dc5ad836b310ff6aeaa4d4826fdffac5ffd500b23147ebd6b3e8d2486: Status 404 returned error can't find the container with id 45c0170dc5ad836b310ff6aeaa4d4826fdffac5ffd500b23147ebd6b3e8d2486 Mar 20 09:02:00 crc kubenswrapper[4823]: I0320 09:02:00.887735 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cbwgk"] Mar 20 09:02:01 crc kubenswrapper[4823]: I0320 09:02:01.316736 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566622-nc9qz" event={"ID":"56782166-f1af-439a-b19b-8eb59caaffe9","Type":"ContainerStarted","Data":"45c0170dc5ad836b310ff6aeaa4d4826fdffac5ffd500b23147ebd6b3e8d2486"} Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.105526 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fafad3e-2369-4997-beb3-c73d884b2276" path="/var/lib/kubelet/pods/4fafad3e-2369-4997-beb3-c73d884b2276/volumes" Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.326923 4823 generic.go:334] "Generic (PLEG): container finished" podID="56782166-f1af-439a-b19b-8eb59caaffe9" containerID="6f4275c05d7506341508bc9cef63cbad8af2db169cebd75b765c564d99f7f698" exitCode=0 Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.327030 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566622-nc9qz" event={"ID":"56782166-f1af-439a-b19b-8eb59caaffe9","Type":"ContainerDied","Data":"6f4275c05d7506341508bc9cef63cbad8af2db169cebd75b765c564d99f7f698"} Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.327934 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cbwgk" podUID="a0ddc452-8cdd-4451-b2a3-215c24347335" containerName="registry-server" containerID="cri-o://72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c" gracePeriod=2 Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.745248 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.818378 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-catalog-content\") pod \"a0ddc452-8cdd-4451-b2a3-215c24347335\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.818435 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-utilities\") pod \"a0ddc452-8cdd-4451-b2a3-215c24347335\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.818475 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5fpf\" (UniqueName: \"kubernetes.io/projected/a0ddc452-8cdd-4451-b2a3-215c24347335-kube-api-access-n5fpf\") pod \"a0ddc452-8cdd-4451-b2a3-215c24347335\" (UID: \"a0ddc452-8cdd-4451-b2a3-215c24347335\") " Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.819206 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-utilities" (OuterVolumeSpecName: "utilities") pod "a0ddc452-8cdd-4451-b2a3-215c24347335" (UID: "a0ddc452-8cdd-4451-b2a3-215c24347335"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.823956 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ddc452-8cdd-4451-b2a3-215c24347335-kube-api-access-n5fpf" (OuterVolumeSpecName: "kube-api-access-n5fpf") pod "a0ddc452-8cdd-4451-b2a3-215c24347335" (UID: "a0ddc452-8cdd-4451-b2a3-215c24347335"). InnerVolumeSpecName "kube-api-access-n5fpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.873590 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0ddc452-8cdd-4451-b2a3-215c24347335" (UID: "a0ddc452-8cdd-4451-b2a3-215c24347335"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.920155 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.920183 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0ddc452-8cdd-4451-b2a3-215c24347335-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:02 crc kubenswrapper[4823]: I0320 09:02:02.920192 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5fpf\" (UniqueName: \"kubernetes.io/projected/a0ddc452-8cdd-4451-b2a3-215c24347335-kube-api-access-n5fpf\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.350660 4823 generic.go:334] "Generic (PLEG): container finished" podID="a0ddc452-8cdd-4451-b2a3-215c24347335" containerID="72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c" exitCode=0 Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.350705 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbwgk" event={"ID":"a0ddc452-8cdd-4451-b2a3-215c24347335","Type":"ContainerDied","Data":"72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c"} Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.350753 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cbwgk" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.350779 4823 scope.go:117] "RemoveContainer" containerID="72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.350765 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbwgk" event={"ID":"a0ddc452-8cdd-4451-b2a3-215c24347335","Type":"ContainerDied","Data":"59ee1ad4b8cd4342b517e4dc68d5641250b4e61319d47c4da45dfb4e98b755a7"} Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.370400 4823 scope.go:117] "RemoveContainer" containerID="de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.376443 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cbwgk"] Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.379859 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cbwgk"] Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.403663 4823 scope.go:117] "RemoveContainer" containerID="a8b057a779abaefe8364a70d4ffea1cb0723b145bb437daef989f1ea146f40ad" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.457394 4823 scope.go:117] "RemoveContainer" containerID="72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c" Mar 20 09:02:03 crc kubenswrapper[4823]: E0320 09:02:03.458082 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c\": container with ID starting with 72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c not found: ID does not exist" containerID="72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.458119 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c"} err="failed to get container status \"72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c\": rpc error: code = NotFound desc = could not find container \"72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c\": container with ID starting with 72bb4e75da531a0b39620b09d0e3e259cebd3ed38ab1d97ea2dbc96a82f4844c not found: ID does not exist" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.458146 4823 scope.go:117] "RemoveContainer" containerID="de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001" Mar 20 09:02:03 crc kubenswrapper[4823]: E0320 09:02:03.458702 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001\": container with ID starting with de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001 not found: ID does not exist" containerID="de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.458773 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001"} err="failed to get container status \"de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001\": rpc error: code = NotFound desc = could not find container \"de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001\": container with ID starting with de88d74361222f83294d8b4fe657fd22a87c08d8b11720a9cee5e4f1f84e4001 not found: ID does not exist" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.458812 4823 scope.go:117] "RemoveContainer" containerID="a8b057a779abaefe8364a70d4ffea1cb0723b145bb437daef989f1ea146f40ad" Mar 20 09:02:03 crc kubenswrapper[4823]: E0320 09:02:03.459337 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8b057a779abaefe8364a70d4ffea1cb0723b145bb437daef989f1ea146f40ad\": container with ID starting with a8b057a779abaefe8364a70d4ffea1cb0723b145bb437daef989f1ea146f40ad not found: ID does not exist" containerID="a8b057a779abaefe8364a70d4ffea1cb0723b145bb437daef989f1ea146f40ad" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.459371 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8b057a779abaefe8364a70d4ffea1cb0723b145bb437daef989f1ea146f40ad"} err="failed to get container status \"a8b057a779abaefe8364a70d4ffea1cb0723b145bb437daef989f1ea146f40ad\": rpc error: code = NotFound desc = could not find container \"a8b057a779abaefe8364a70d4ffea1cb0723b145bb437daef989f1ea146f40ad\": container with ID starting with a8b057a779abaefe8364a70d4ffea1cb0723b145bb437daef989f1ea146f40ad not found: ID does not exist" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.668719 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566622-nc9qz" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.849560 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x448\" (UniqueName: \"kubernetes.io/projected/56782166-f1af-439a-b19b-8eb59caaffe9-kube-api-access-8x448\") pod \"56782166-f1af-439a-b19b-8eb59caaffe9\" (UID: \"56782166-f1af-439a-b19b-8eb59caaffe9\") " Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.873901 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56782166-f1af-439a-b19b-8eb59caaffe9-kube-api-access-8x448" (OuterVolumeSpecName: "kube-api-access-8x448") pod "56782166-f1af-439a-b19b-8eb59caaffe9" (UID: "56782166-f1af-439a-b19b-8eb59caaffe9"). InnerVolumeSpecName "kube-api-access-8x448". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:03 crc kubenswrapper[4823]: I0320 09:02:03.951642 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x448\" (UniqueName: \"kubernetes.io/projected/56782166-f1af-439a-b19b-8eb59caaffe9-kube-api-access-8x448\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:04 crc kubenswrapper[4823]: I0320 09:02:04.110614 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0ddc452-8cdd-4451-b2a3-215c24347335" path="/var/lib/kubelet/pods/a0ddc452-8cdd-4451-b2a3-215c24347335/volumes" Mar 20 09:02:04 crc kubenswrapper[4823]: I0320 09:02:04.358448 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566622-nc9qz" event={"ID":"56782166-f1af-439a-b19b-8eb59caaffe9","Type":"ContainerDied","Data":"45c0170dc5ad836b310ff6aeaa4d4826fdffac5ffd500b23147ebd6b3e8d2486"} Mar 20 09:02:04 crc kubenswrapper[4823]: I0320 09:02:04.358507 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45c0170dc5ad836b310ff6aeaa4d4826fdffac5ffd500b23147ebd6b3e8d2486" Mar 20 09:02:04 crc kubenswrapper[4823]: I0320 09:02:04.358531 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566622-nc9qz" Mar 20 09:02:09 crc kubenswrapper[4823]: I0320 09:02:09.429963 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:02:09 crc kubenswrapper[4823]: I0320 09:02:09.430035 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:02:09 crc kubenswrapper[4823]: I0320 09:02:09.430086 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:02:09 crc kubenswrapper[4823]: I0320 09:02:09.430657 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:02:09 crc kubenswrapper[4823]: I0320 09:02:09.430712 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb" gracePeriod=600 Mar 20 09:02:09 crc kubenswrapper[4823]: I0320 09:02:09.463678 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hlgq9"] Mar 20 09:02:10 crc kubenswrapper[4823]: I0320 09:02:10.401550 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb" exitCode=0 Mar 20 09:02:10 crc kubenswrapper[4823]: I0320 09:02:10.401589 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb"} Mar 20 09:02:10 crc kubenswrapper[4823]: I0320 09:02:10.401839 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"6d87d616f40330266a80a533f4cd60db866a189b08f07fd958207c63263d156f"} Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.110397 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2"] Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.110959 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" containerName="controller-manager" containerID="cri-o://8fa7b75a76a8e3f3c0dc367d0bf6dc097b3a9526fe7cd13738a217d3b18882ad" gracePeriod=30 Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.245664 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7"] Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.245880 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" containerName="route-controller-manager" containerID="cri-o://99ed05a7e22440487990806107359d6a57e851f13d011a823103b233707c4569" gracePeriod=30 Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.455539 4823 generic.go:334] "Generic (PLEG): container finished" podID="ddd258bb-2c2e-4211-94b3-947051b719f0" containerID="8fa7b75a76a8e3f3c0dc367d0bf6dc097b3a9526fe7cd13738a217d3b18882ad" exitCode=0 Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.455629 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" event={"ID":"ddd258bb-2c2e-4211-94b3-947051b719f0","Type":"ContainerDied","Data":"8fa7b75a76a8e3f3c0dc367d0bf6dc097b3a9526fe7cd13738a217d3b18882ad"} Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.457279 4823 generic.go:334] "Generic (PLEG): container finished" podID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" containerID="99ed05a7e22440487990806107359d6a57e851f13d011a823103b233707c4569" exitCode=0 Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.457330 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" event={"ID":"2fca0d9b-55e0-49ae-a768-b2d1c19f1635","Type":"ContainerDied","Data":"99ed05a7e22440487990806107359d6a57e851f13d011a823103b233707c4569"} Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.697521 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.701899 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.840122 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-client-ca\") pod \"ddd258bb-2c2e-4211-94b3-947051b719f0\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.840429 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-config\") pod \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.840454 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-client-ca\") pod \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.840529 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b77pd\" (UniqueName: \"kubernetes.io/projected/ddd258bb-2c2e-4211-94b3-947051b719f0-kube-api-access-b77pd\") pod \"ddd258bb-2c2e-4211-94b3-947051b719f0\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.840550 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd7pd\" (UniqueName: \"kubernetes.io/projected/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-kube-api-access-jd7pd\") pod \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.840599 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-serving-cert\") pod \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\" (UID: \"2fca0d9b-55e0-49ae-a768-b2d1c19f1635\") " Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.840622 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddd258bb-2c2e-4211-94b3-947051b719f0-serving-cert\") pod \"ddd258bb-2c2e-4211-94b3-947051b719f0\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.840658 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-config\") pod \"ddd258bb-2c2e-4211-94b3-947051b719f0\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.840694 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-proxy-ca-bundles\") pod \"ddd258bb-2c2e-4211-94b3-947051b719f0\" (UID: \"ddd258bb-2c2e-4211-94b3-947051b719f0\") " Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.841763 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ddd258bb-2c2e-4211-94b3-947051b719f0" (UID: "ddd258bb-2c2e-4211-94b3-947051b719f0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.842254 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-client-ca" (OuterVolumeSpecName: "client-ca") pod "ddd258bb-2c2e-4211-94b3-947051b719f0" (UID: "ddd258bb-2c2e-4211-94b3-947051b719f0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.842946 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-config" (OuterVolumeSpecName: "config") pod "2fca0d9b-55e0-49ae-a768-b2d1c19f1635" (UID: "2fca0d9b-55e0-49ae-a768-b2d1c19f1635"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.843468 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-client-ca" (OuterVolumeSpecName: "client-ca") pod "2fca0d9b-55e0-49ae-a768-b2d1c19f1635" (UID: "2fca0d9b-55e0-49ae-a768-b2d1c19f1635"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.846622 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-config" (OuterVolumeSpecName: "config") pod "ddd258bb-2c2e-4211-94b3-947051b719f0" (UID: "ddd258bb-2c2e-4211-94b3-947051b719f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.848722 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddd258bb-2c2e-4211-94b3-947051b719f0-kube-api-access-b77pd" (OuterVolumeSpecName: "kube-api-access-b77pd") pod "ddd258bb-2c2e-4211-94b3-947051b719f0" (UID: "ddd258bb-2c2e-4211-94b3-947051b719f0"). InnerVolumeSpecName "kube-api-access-b77pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.848829 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2fca0d9b-55e0-49ae-a768-b2d1c19f1635" (UID: "2fca0d9b-55e0-49ae-a768-b2d1c19f1635"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.848940 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-kube-api-access-jd7pd" (OuterVolumeSpecName: "kube-api-access-jd7pd") pod "2fca0d9b-55e0-49ae-a768-b2d1c19f1635" (UID: "2fca0d9b-55e0-49ae-a768-b2d1c19f1635"). InnerVolumeSpecName "kube-api-access-jd7pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.849791 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd258bb-2c2e-4211-94b3-947051b719f0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ddd258bb-2c2e-4211-94b3-947051b719f0" (UID: "ddd258bb-2c2e-4211-94b3-947051b719f0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.942273 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b77pd\" (UniqueName: \"kubernetes.io/projected/ddd258bb-2c2e-4211-94b3-947051b719f0-kube-api-access-b77pd\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.942308 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd7pd\" (UniqueName: \"kubernetes.io/projected/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-kube-api-access-jd7pd\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.942320 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.942329 4823 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddd258bb-2c2e-4211-94b3-947051b719f0-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.942337 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.942348 4823 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.942355 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddd258bb-2c2e-4211-94b3-947051b719f0-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.942363 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:17 crc kubenswrapper[4823]: I0320 09:02:17.942370 4823 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fca0d9b-55e0-49ae-a768-b2d1c19f1635-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.088393 4823 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.088579 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ddc452-8cdd-4451-b2a3-215c24347335" containerName="extract-content" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.088592 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ddc452-8cdd-4451-b2a3-215c24347335" containerName="extract-content" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.088604 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ddc452-8cdd-4451-b2a3-215c24347335" containerName="registry-server" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.088610 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ddc452-8cdd-4451-b2a3-215c24347335" containerName="registry-server" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.088620 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56782166-f1af-439a-b19b-8eb59caaffe9" containerName="oc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.088626 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="56782166-f1af-439a-b19b-8eb59caaffe9" containerName="oc" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.088634 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" containerName="controller-manager" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.088640 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" containerName="controller-manager" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.088652 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ddc452-8cdd-4451-b2a3-215c24347335" containerName="extract-utilities" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.088658 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ddc452-8cdd-4451-b2a3-215c24347335" containerName="extract-utilities" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.088668 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" containerName="route-controller-manager" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.088674 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" containerName="route-controller-manager" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.088774 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" containerName="controller-manager" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.088783 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ddc452-8cdd-4451-b2a3-215c24347335" containerName="registry-server" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.088796 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="56782166-f1af-439a-b19b-8eb59caaffe9" containerName="oc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.088803 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" containerName="route-controller-manager" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.089058 4823 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.089333 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976" gracePeriod=15 Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.089366 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89" gracePeriod=15 Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.089397 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e" gracePeriod=15 Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.089312 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a" gracePeriod=15 Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.089458 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8" gracePeriod=15 Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.089608 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091463 4823 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.091649 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091669 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.091681 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091689 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.091700 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091708 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.091718 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091725 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.091753 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091762 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.091780 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091788 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.091797 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091804 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.091812 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091820 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.091833 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091841 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091958 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091971 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091981 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.091995 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.092004 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.092015 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.092027 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.092131 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.092141 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.092270 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.092482 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.130157 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.245712 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.246046 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.246074 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.246109 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.246133 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.246235 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.246270 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.246301 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.346884 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.346928 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.346956 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.346995 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347023 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347073 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347094 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347122 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347209 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347251 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347279 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347307 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347334 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347362 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347400 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.347422 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.428477 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:18 crc kubenswrapper[4823]: W0320 09:02:18.457029 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-a07f77053559ba7cbb37cb50dc4832550ae838c8d84658c74d6ff8b45fccdec4 WatchSource:0}: Error finding container a07f77053559ba7cbb37cb50dc4832550ae838c8d84658c74d6ff8b45fccdec4: Status 404 returned error can't find the container with id a07f77053559ba7cbb37cb50dc4832550ae838c8d84658c74d6ff8b45fccdec4 Mar 20 09:02:18 crc kubenswrapper[4823]: E0320 09:02:18.460407 4823 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189e812e27627222 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 09:02:18.459796002 +0000 UTC m=+305.145692376,LastTimestamp:2026-03-20 09:02:18.459796002 +0000 UTC m=+305.145692376,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.469054 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.469134 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" event={"ID":"2fca0d9b-55e0-49ae-a768-b2d1c19f1635","Type":"ContainerDied","Data":"fff6ac43622a18efe15cfddc42e0fa727a34af6d1c90ba28888593eb67d2d7f6"} Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.469278 4823 scope.go:117] "RemoveContainer" containerID="99ed05a7e22440487990806107359d6a57e851f13d011a823103b233707c4569" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.470687 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.471796 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" event={"ID":"ddd258bb-2c2e-4211-94b3-947051b719f0","Type":"ContainerDied","Data":"7d93edb4ac6fde0ecf3e0b5b477b5d314633253a6877fab6c30955eb4a38846c"} Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.471915 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.475078 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.475563 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.476355 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.476614 4823 generic.go:334] "Generic (PLEG): container finished" podID="c826ac6f-8206-4ce3-a59a-c03afefbc038" containerID="a9885ecfdcf3152d8ac207eea93d988198d761502409927563c405c09edbd829" exitCode=0 Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.476645 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c826ac6f-8206-4ce3-a59a-c03afefbc038","Type":"ContainerDied","Data":"a9885ecfdcf3152d8ac207eea93d988198d761502409927563c405c09edbd829"} Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.476887 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.477237 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.477457 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.478065 4823 status_manager.go:851] "Failed to get status for pod" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.478811 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a07f77053559ba7cbb37cb50dc4832550ae838c8d84658c74d6ff8b45fccdec4"} Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.478944 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.482890 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.485046 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.487722 4823 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976" exitCode=0 Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.487818 4823 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8" exitCode=0 Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.487827 4823 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e" exitCode=0 Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.487833 4823 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89" exitCode=2 Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.521883 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.522361 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.522651 4823 status_manager.go:851] "Failed to get status for pod" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.524013 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.524533 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.525373 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.525639 4823 status_manager.go:851] "Failed to get status for pod" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.525918 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.528497 4823 scope.go:117] "RemoveContainer" containerID="8fa7b75a76a8e3f3c0dc367d0bf6dc097b3a9526fe7cd13738a217d3b18882ad" Mar 20 09:02:18 crc kubenswrapper[4823]: I0320 09:02:18.546135 4823 scope.go:117] "RemoveContainer" containerID="6f166fae2bf61c8da6671628ce2ba0ced8d049a40476b3660385c0847dbe0099" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.499591 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490"} Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.500528 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.501000 4823 status_manager.go:851] "Failed to get status for pod" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.501470 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.501875 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.504058 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.808464 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.809620 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.810179 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.810503 4823 status_manager.go:851] "Failed to get status for pod" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.810891 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.970824 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c826ac6f-8206-4ce3-a59a-c03afefbc038-kube-api-access\") pod \"c826ac6f-8206-4ce3-a59a-c03afefbc038\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.970903 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-kubelet-dir\") pod \"c826ac6f-8206-4ce3-a59a-c03afefbc038\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.970955 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-var-lock\") pod \"c826ac6f-8206-4ce3-a59a-c03afefbc038\" (UID: \"c826ac6f-8206-4ce3-a59a-c03afefbc038\") " Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.971365 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c826ac6f-8206-4ce3-a59a-c03afefbc038" (UID: "c826ac6f-8206-4ce3-a59a-c03afefbc038"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.971479 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-var-lock" (OuterVolumeSpecName: "var-lock") pod "c826ac6f-8206-4ce3-a59a-c03afefbc038" (UID: "c826ac6f-8206-4ce3-a59a-c03afefbc038"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.971581 4823 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:19 crc kubenswrapper[4823]: I0320 09:02:19.977958 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c826ac6f-8206-4ce3-a59a-c03afefbc038-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c826ac6f-8206-4ce3-a59a-c03afefbc038" (UID: "c826ac6f-8206-4ce3-a59a-c03afefbc038"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.073213 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c826ac6f-8206-4ce3-a59a-c03afefbc038-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.073551 4823 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c826ac6f-8206-4ce3-a59a-c03afefbc038-var-lock\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.480570 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.481983 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.482626 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.483350 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.483969 4823 status_manager.go:851] "Failed to get status for pod" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.484499 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.484954 4823 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.517555 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.517585 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c826ac6f-8206-4ce3-a59a-c03afefbc038","Type":"ContainerDied","Data":"1fbe640b66ab8c698502de38b9d8a1e61cf48c9e6ef66730b223fa8b0d05509b"} Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.517803 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fbe640b66ab8c698502de38b9d8a1e61cf48c9e6ef66730b223fa8b0d05509b" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.522855 4823 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.523469 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.524117 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.524827 4823 status_manager.go:851] "Failed to get status for pod" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.525047 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.525408 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.526395 4823 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a" exitCode=0 Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.526473 4823 scope.go:117] "RemoveContainer" containerID="b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.526598 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.552566 4823 scope.go:117] "RemoveContainer" containerID="7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.572589 4823 scope.go:117] "RemoveContainer" containerID="9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.582230 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.582354 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.582371 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.582450 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.582500 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.582568 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.582883 4823 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.582916 4823 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.582940 4823 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.596339 4823 scope.go:117] "RemoveContainer" containerID="ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.616709 4823 scope.go:117] "RemoveContainer" containerID="f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.640664 4823 scope.go:117] "RemoveContainer" containerID="69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.675304 4823 scope.go:117] "RemoveContainer" containerID="b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976" Mar 20 09:02:20 crc kubenswrapper[4823]: E0320 09:02:20.676268 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\": container with ID starting with b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976 not found: ID does not exist" containerID="b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.676355 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976"} err="failed to get container status \"b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\": rpc error: code = NotFound desc = could not find container \"b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976\": container with ID starting with b37eb278ada55c59a0336bda71b6d937db1e4e3e7a508b3e473cad1312fe8976 not found: ID does not exist" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.676416 4823 scope.go:117] "RemoveContainer" containerID="7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8" Mar 20 09:02:20 crc kubenswrapper[4823]: E0320 09:02:20.677138 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\": container with ID starting with 7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8 not found: ID does not exist" containerID="7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.677210 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8"} err="failed to get container status \"7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\": rpc error: code = NotFound desc = could not find container \"7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8\": container with ID starting with 7d5bbe529e8d7a8a6a2e2c8210ba41096aada35b09eaf1b0b9f4bcaaf33ea2e8 not found: ID does not exist" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.677251 4823 scope.go:117] "RemoveContainer" containerID="9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e" Mar 20 09:02:20 crc kubenswrapper[4823]: E0320 09:02:20.677707 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\": container with ID starting with 9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e not found: ID does not exist" containerID="9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.677825 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e"} err="failed to get container status \"9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\": rpc error: code = NotFound desc = could not find container \"9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e\": container with ID starting with 9e5e5b07f7792d44abf604242eacc6fab7958640c013658a71d0283a8bbe370e not found: ID does not exist" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.677863 4823 scope.go:117] "RemoveContainer" containerID="ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89" Mar 20 09:02:20 crc kubenswrapper[4823]: E0320 09:02:20.678727 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\": container with ID starting with ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89 not found: ID does not exist" containerID="ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.678814 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89"} err="failed to get container status \"ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\": rpc error: code = NotFound desc = could not find container \"ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89\": container with ID starting with ddab6bdaa619b1c18f0db361550e1280502784b5dc3b4f909b89bb8554345c89 not found: ID does not exist" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.678854 4823 scope.go:117] "RemoveContainer" containerID="f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a" Mar 20 09:02:20 crc kubenswrapper[4823]: E0320 09:02:20.679689 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\": container with ID starting with f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a not found: ID does not exist" containerID="f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.679780 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a"} err="failed to get container status \"f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\": rpc error: code = NotFound desc = could not find container \"f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a\": container with ID starting with f55a17847b4639f1fc589d8d3f4eeca766f0c49347c3bbec0521cf005d7a4e3a not found: ID does not exist" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.679823 4823 scope.go:117] "RemoveContainer" containerID="69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5" Mar 20 09:02:20 crc kubenswrapper[4823]: E0320 09:02:20.680629 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\": container with ID starting with 69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5 not found: ID does not exist" containerID="69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.680693 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5"} err="failed to get container status \"69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\": rpc error: code = NotFound desc = could not find container \"69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5\": container with ID starting with 69ef3ed983869da20b5006afd79eb3f42282e0ccf1c968b144c7f4633a3fb1e5 not found: ID does not exist" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.853710 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.854314 4823 status_manager.go:851] "Failed to get status for pod" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.854850 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.855395 4823 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:20 crc kubenswrapper[4823]: I0320 09:02:20.856111 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:22 crc kubenswrapper[4823]: I0320 09:02:22.108157 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 20 09:02:22 crc kubenswrapper[4823]: E0320 09:02:22.292477 4823 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189e812e27627222 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 09:02:18.459796002 +0000 UTC m=+305.145692376,LastTimestamp:2026-03-20 09:02:18.459796002 +0000 UTC m=+305.145692376,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 09:02:24 crc kubenswrapper[4823]: I0320 09:02:24.102066 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:24 crc kubenswrapper[4823]: I0320 09:02:24.102910 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:24 crc kubenswrapper[4823]: I0320 09:02:24.103250 4823 status_manager.go:851] "Failed to get status for pod" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:24 crc kubenswrapper[4823]: I0320 09:02:24.103550 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:27 crc kubenswrapper[4823]: E0320 09:02:27.894731 4823 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:27 crc kubenswrapper[4823]: E0320 09:02:27.896896 4823 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:27 crc kubenswrapper[4823]: E0320 09:02:27.897391 4823 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:27 crc kubenswrapper[4823]: E0320 09:02:27.897888 4823 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:27 crc kubenswrapper[4823]: E0320 09:02:27.898314 4823 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:27 crc kubenswrapper[4823]: I0320 09:02:27.898355 4823 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 20 09:02:27 crc kubenswrapper[4823]: E0320 09:02:27.898713 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="200ms" Mar 20 09:02:28 crc kubenswrapper[4823]: E0320 09:02:28.099850 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="400ms" Mar 20 09:02:28 crc kubenswrapper[4823]: E0320 09:02:28.500607 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="800ms" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.097370 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.099002 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.103649 4823 status_manager.go:851] "Failed to get status for pod" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.105857 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.106269 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.116169 4823 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.116201 4823 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:29 crc kubenswrapper[4823]: E0320 09:02:29.116560 4823 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.117097 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:29 crc kubenswrapper[4823]: W0320 09:02:29.151632 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-e0d38e5200e7bfcd8355ee629ffec2aa1ec00ad7e96022e8f853855aba76a30a WatchSource:0}: Error finding container e0d38e5200e7bfcd8355ee629ffec2aa1ec00ad7e96022e8f853855aba76a30a: Status 404 returned error can't find the container with id e0d38e5200e7bfcd8355ee629ffec2aa1ec00ad7e96022e8f853855aba76a30a Mar 20 09:02:29 crc kubenswrapper[4823]: E0320 09:02:29.302585 4823 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="1.6s" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.600055 4823 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="ae4b453c494a0293060c419c9094d12a938dfa9163feba79865616e9737405f3" exitCode=0 Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.600114 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"ae4b453c494a0293060c419c9094d12a938dfa9163feba79865616e9737405f3"} Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.600151 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e0d38e5200e7bfcd8355ee629ffec2aa1ec00ad7e96022e8f853855aba76a30a"} Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.600481 4823 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.600510 4823 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.601026 4823 status_manager.go:851] "Failed to get status for pod" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" pod="openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6bffb6f9f6-4sqz2\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: E0320 09:02:29.601042 4823 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.601495 4823 status_manager.go:851] "Failed to get status for pod" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" pod="openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5c54d7bcb-p4xv7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.601928 4823 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: I0320 09:02:29.602269 4823 status_manager.go:851] "Failed to get status for pod" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: E0320 09:02:29.863516 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:02:29Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:02:29Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:02:29Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T09:02:29Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: E0320 09:02:29.867916 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: E0320 09:02:29.868429 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: E0320 09:02:29.868911 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: E0320 09:02:29.869285 4823 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 20 09:02:29 crc kubenswrapper[4823]: E0320 09:02:29.869311 4823 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 09:02:30 crc kubenswrapper[4823]: I0320 09:02:30.611559 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e462623c143019b4feefe51f95f9a58d95ee4a2d6834e6f5b12560a1b8a953fd"} Mar 20 09:02:30 crc kubenswrapper[4823]: I0320 09:02:30.611605 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1ae1b76af7ddcf56b7ed791a1039201db758f78e225592a693309671eac2abcc"} Mar 20 09:02:30 crc kubenswrapper[4823]: I0320 09:02:30.611615 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"964d12537ea9332b1baf06252278a61a741d64c1db70afa4c1ee4ab0b003c6e4"} Mar 20 09:02:30 crc kubenswrapper[4823]: I0320 09:02:30.611625 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"29fae688b7f6f2de7cea111ac5d3df4323925d94a935c3166ae5a1b2b67428b2"} Mar 20 09:02:31 crc kubenswrapper[4823]: I0320 09:02:31.624237 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"03103f566bd0d31e70a0f2cded3e7bba386c142770c32ee816fe57fd5c31284c"} Mar 20 09:02:31 crc kubenswrapper[4823]: I0320 09:02:31.624636 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:31 crc kubenswrapper[4823]: I0320 09:02:31.624467 4823 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:31 crc kubenswrapper[4823]: I0320 09:02:31.624684 4823 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:32 crc kubenswrapper[4823]: I0320 09:02:32.645184 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 09:02:32 crc kubenswrapper[4823]: I0320 09:02:32.646218 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 20 09:02:32 crc kubenswrapper[4823]: I0320 09:02:32.646297 4823 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483" exitCode=1 Mar 20 09:02:32 crc kubenswrapper[4823]: I0320 09:02:32.646342 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483"} Mar 20 09:02:32 crc kubenswrapper[4823]: I0320 09:02:32.647063 4823 scope.go:117] "RemoveContainer" containerID="2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483" Mar 20 09:02:33 crc kubenswrapper[4823]: I0320 09:02:33.655940 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 09:02:33 crc kubenswrapper[4823]: I0320 09:02:33.657271 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 20 09:02:33 crc kubenswrapper[4823]: I0320 09:02:33.657351 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"19b0d3a42e7bec22ce3d90fee067806f5e1a58ddfd40e38ea8c13a134ff093e1"} Mar 20 09:02:34 crc kubenswrapper[4823]: I0320 09:02:34.117789 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:34 crc kubenswrapper[4823]: I0320 09:02:34.117861 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:34 crc kubenswrapper[4823]: I0320 09:02:34.125546 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:34 crc kubenswrapper[4823]: I0320 09:02:34.503868 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" podUID="66cb5542-0914-40bf-b27d-115a56b07012" containerName="oauth-openshift" containerID="cri-o://b521b5eb4d29072d233223caf224b844651693501345b5c86f510b27a5cd6a8d" gracePeriod=15 Mar 20 09:02:34 crc kubenswrapper[4823]: I0320 09:02:34.665614 4823 generic.go:334] "Generic (PLEG): container finished" podID="66cb5542-0914-40bf-b27d-115a56b07012" containerID="b521b5eb4d29072d233223caf224b844651693501345b5c86f510b27a5cd6a8d" exitCode=0 Mar 20 09:02:34 crc kubenswrapper[4823]: I0320 09:02:34.665694 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" event={"ID":"66cb5542-0914-40bf-b27d-115a56b07012","Type":"ContainerDied","Data":"b521b5eb4d29072d233223caf224b844651693501345b5c86f510b27a5cd6a8d"} Mar 20 09:02:34 crc kubenswrapper[4823]: I0320 09:02:34.934383 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.068373 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-audit-policies\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.068426 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-cliconfig\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.068452 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66cb5542-0914-40bf-b27d-115a56b07012-audit-dir\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.068650 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66cb5542-0914-40bf-b27d-115a56b07012-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.069066 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.069077 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.069186 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-login\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.069769 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.070290 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-service-ca\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.070337 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-ocp-branding-template\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.070363 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-serving-cert\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.070383 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-idp-0-file-data\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.070407 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2r2b\" (UniqueName: \"kubernetes.io/projected/66cb5542-0914-40bf-b27d-115a56b07012-kube-api-access-b2r2b\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.070442 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-trusted-ca-bundle\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.070463 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-provider-selection\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.070754 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-error\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.070798 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-session\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.070830 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-router-certs\") pod \"66cb5542-0914-40bf-b27d-115a56b07012\" (UID: \"66cb5542-0914-40bf-b27d-115a56b07012\") " Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.071142 4823 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.071169 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.071183 4823 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66cb5542-0914-40bf-b27d-115a56b07012-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.071197 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.071775 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.075145 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.075661 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.075820 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66cb5542-0914-40bf-b27d-115a56b07012-kube-api-access-b2r2b" (OuterVolumeSpecName: "kube-api-access-b2r2b") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "kube-api-access-b2r2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.076336 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.076498 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.076624 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.082941 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.083075 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.083236 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "66cb5542-0914-40bf-b27d-115a56b07012" (UID: "66cb5542-0914-40bf-b27d-115a56b07012"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.172024 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.172063 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.172078 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.172090 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.172103 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.172116 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.172128 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.172139 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.172151 4823 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66cb5542-0914-40bf-b27d-115a56b07012-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.172163 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2r2b\" (UniqueName: \"kubernetes.io/projected/66cb5542-0914-40bf-b27d-115a56b07012-kube-api-access-b2r2b\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.673942 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" event={"ID":"66cb5542-0914-40bf-b27d-115a56b07012","Type":"ContainerDied","Data":"80ab244d539acc8e64fddf414d68a41c53f520ad8ef72603dfed045bc3de3088"} Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.674028 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hlgq9" Mar 20 09:02:35 crc kubenswrapper[4823]: I0320 09:02:35.674046 4823 scope.go:117] "RemoveContainer" containerID="b521b5eb4d29072d233223caf224b844651693501345b5c86f510b27a5cd6a8d" Mar 20 09:02:36 crc kubenswrapper[4823]: I0320 09:02:36.272223 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 09:02:36 crc kubenswrapper[4823]: I0320 09:02:36.634627 4823 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:36 crc kubenswrapper[4823]: I0320 09:02:36.681082 4823 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:36 crc kubenswrapper[4823]: I0320 09:02:36.681117 4823 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:36 crc kubenswrapper[4823]: I0320 09:02:36.686863 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:36 crc kubenswrapper[4823]: I0320 09:02:36.738920 4823 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a1944220-f32c-4bc8-833c-f4fd896fb48d" Mar 20 09:02:37 crc kubenswrapper[4823]: I0320 09:02:37.685339 4823 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:37 crc kubenswrapper[4823]: I0320 09:02:37.685730 4823 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:37 crc kubenswrapper[4823]: I0320 09:02:37.689596 4823 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a1944220-f32c-4bc8-833c-f4fd896fb48d" Mar 20 09:02:41 crc kubenswrapper[4823]: I0320 09:02:41.572258 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 09:02:41 crc kubenswrapper[4823]: I0320 09:02:41.572868 4823 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 20 09:02:41 crc kubenswrapper[4823]: I0320 09:02:41.572958 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 20 09:02:42 crc kubenswrapper[4823]: I0320 09:02:42.611986 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 20 09:02:42 crc kubenswrapper[4823]: I0320 09:02:42.721705 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 20 09:02:43 crc kubenswrapper[4823]: I0320 09:02:43.154620 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 20 09:02:43 crc kubenswrapper[4823]: I0320 09:02:43.193184 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 20 09:02:43 crc kubenswrapper[4823]: I0320 09:02:43.323450 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 20 09:02:43 crc kubenswrapper[4823]: I0320 09:02:43.544885 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 20 09:02:44 crc kubenswrapper[4823]: I0320 09:02:44.122190 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 20 09:02:45 crc kubenswrapper[4823]: I0320 09:02:45.011436 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 20 09:02:45 crc kubenswrapper[4823]: I0320 09:02:45.528799 4823 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 20 09:02:46 crc kubenswrapper[4823]: I0320 09:02:46.045374 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 20 09:02:46 crc kubenswrapper[4823]: I0320 09:02:46.170237 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 20 09:02:46 crc kubenswrapper[4823]: I0320 09:02:46.177795 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 20 09:02:46 crc kubenswrapper[4823]: I0320 09:02:46.192633 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 20 09:02:46 crc kubenswrapper[4823]: I0320 09:02:46.681499 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 20 09:02:46 crc kubenswrapper[4823]: I0320 09:02:46.681499 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 20 09:02:46 crc kubenswrapper[4823]: I0320 09:02:46.684871 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 20 09:02:46 crc kubenswrapper[4823]: I0320 09:02:46.929293 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 20 09:02:47 crc kubenswrapper[4823]: I0320 09:02:47.210778 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 20 09:02:48 crc kubenswrapper[4823]: I0320 09:02:48.566396 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 20 09:02:48 crc kubenswrapper[4823]: I0320 09:02:48.630370 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 20 09:02:48 crc kubenswrapper[4823]: I0320 09:02:48.973504 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 20 09:02:49 crc kubenswrapper[4823]: I0320 09:02:49.185193 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 20 09:02:49 crc kubenswrapper[4823]: I0320 09:02:49.247550 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 20 09:02:49 crc kubenswrapper[4823]: I0320 09:02:49.329130 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 20 09:02:49 crc kubenswrapper[4823]: I0320 09:02:49.597039 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 20 09:02:49 crc kubenswrapper[4823]: I0320 09:02:49.845066 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 20 09:02:50 crc kubenswrapper[4823]: I0320 09:02:50.014287 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 20 09:02:50 crc kubenswrapper[4823]: I0320 09:02:50.035896 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 20 09:02:50 crc kubenswrapper[4823]: I0320 09:02:50.422212 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 20 09:02:50 crc kubenswrapper[4823]: I0320 09:02:50.425659 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 20 09:02:50 crc kubenswrapper[4823]: I0320 09:02:50.440108 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 20 09:02:50 crc kubenswrapper[4823]: I0320 09:02:50.502080 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 20 09:02:50 crc kubenswrapper[4823]: I0320 09:02:50.613548 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 20 09:02:50 crc kubenswrapper[4823]: I0320 09:02:50.691700 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 20 09:02:50 crc kubenswrapper[4823]: I0320 09:02:50.942224 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 20 09:02:50 crc kubenswrapper[4823]: I0320 09:02:50.952946 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 20 09:02:51 crc kubenswrapper[4823]: I0320 09:02:51.027431 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 20 09:02:51 crc kubenswrapper[4823]: I0320 09:02:51.154936 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 20 09:02:51 crc kubenswrapper[4823]: I0320 09:02:51.213232 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 20 09:02:51 crc kubenswrapper[4823]: I0320 09:02:51.304688 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 20 09:02:51 crc kubenswrapper[4823]: I0320 09:02:51.573814 4823 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 20 09:02:51 crc kubenswrapper[4823]: I0320 09:02:51.573904 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 20 09:02:51 crc kubenswrapper[4823]: I0320 09:02:51.616426 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 20 09:02:51 crc kubenswrapper[4823]: I0320 09:02:51.618771 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 20 09:02:51 crc kubenswrapper[4823]: I0320 09:02:51.851945 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 20 09:02:51 crc kubenswrapper[4823]: I0320 09:02:51.923815 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.225691 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.443727 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.553609 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.677038 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.706653 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.748833 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.760830 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.777948 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.834132 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.835786 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.903855 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.905247 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 20 09:02:52 crc kubenswrapper[4823]: I0320 09:02:52.968404 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.035945 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.157603 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.224554 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.228489 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.255788 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.289124 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.359216 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.446601 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.465062 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.481473 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.532673 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.619942 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.667519 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.784907 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.840441 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.841847 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.937598 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 20 09:02:53 crc kubenswrapper[4823]: I0320 09:02:53.954682 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.180989 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.273810 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.357275 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.559600 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.567951 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.572110 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.607449 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.608086 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.611374 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.660070 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.678492 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.704183 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.764238 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.806499 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.823060 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 20 09:02:54 crc kubenswrapper[4823]: I0320 09:02:54.955894 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.268887 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.306037 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.365033 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.408898 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.418780 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.457373 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.477194 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.545361 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.561146 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.610793 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.763167 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.795494 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.804810 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.811462 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.816904 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.888463 4823 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 20 09:02:55 crc kubenswrapper[4823]: I0320 09:02:55.954013 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 20 09:02:56 crc kubenswrapper[4823]: I0320 09:02:56.031342 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 20 09:02:56 crc kubenswrapper[4823]: I0320 09:02:56.290112 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 20 09:02:56 crc kubenswrapper[4823]: I0320 09:02:56.402455 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 20 09:02:56 crc kubenswrapper[4823]: I0320 09:02:56.415493 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 20 09:02:56 crc kubenswrapper[4823]: I0320 09:02:56.515780 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 20 09:02:56 crc kubenswrapper[4823]: I0320 09:02:56.532182 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 20 09:02:56 crc kubenswrapper[4823]: I0320 09:02:56.753919 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 20 09:02:56 crc kubenswrapper[4823]: I0320 09:02:56.864104 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 20 09:02:56 crc kubenswrapper[4823]: I0320 09:02:56.867967 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 20 09:02:56 crc kubenswrapper[4823]: I0320 09:02:56.881495 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 20 09:02:56 crc kubenswrapper[4823]: I0320 09:02:56.968057 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.021918 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.080999 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.093210 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.141926 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.141929 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.265099 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.271795 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.398934 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.408504 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.436641 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.453761 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.467648 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.592917 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.666657 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.729198 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.753235 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.836958 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.839871 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.914332 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 20 09:02:57 crc kubenswrapper[4823]: I0320 09:02:57.983029 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.025381 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.121272 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.181194 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.283413 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.298878 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.433272 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.592338 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.612807 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.660608 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.690988 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.722106 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.786567 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.818453 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.842479 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.892992 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 20 09:02:58 crc kubenswrapper[4823]: I0320 09:02:58.974763 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.005005 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.167839 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.178924 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.211931 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.281021 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.297196 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.403222 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.426209 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.438692 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.438780 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.728153 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.730603 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.852435 4823 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.854625 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.856337 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=41.856315822 podStartE2EDuration="41.856315822s" podCreationTimestamp="2026-03-20 09:02:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:02:36.652900567 +0000 UTC m=+323.338796941" watchObservedRunningTime="2026-03-20 09:02:59.856315822 +0000 UTC m=+346.542212226" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.857483 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.861072 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c54d7bcb-p4xv7","openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-hlgq9","openshift-controller-manager/controller-manager-6bffb6f9f6-4sqz2"] Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.861173 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-965c65cc7-q6vgf","openshift-kube-apiserver/kube-apiserver-crc","openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s"] Mar 20 09:02:59 crc kubenswrapper[4823]: E0320 09:02:59.861442 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" containerName="installer" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.861469 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" containerName="installer" Mar 20 09:02:59 crc kubenswrapper[4823]: E0320 09:02:59.861493 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66cb5542-0914-40bf-b27d-115a56b07012" containerName="oauth-openshift" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.861507 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="66cb5542-0914-40bf-b27d-115a56b07012" containerName="oauth-openshift" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.861712 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="c826ac6f-8206-4ce3-a59a-c03afefbc038" containerName="installer" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.861772 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="66cb5542-0914-40bf-b27d-115a56b07012" containerName="oauth-openshift" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.861856 4823 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.861922 4823 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1d9b9db9-41b1-4510-bfe7-f57b07b66749" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.862618 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.863681 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.869733 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.874904 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.875429 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.876337 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.876339 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.876448 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.876602 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.876828 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.876858 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.877057 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.877681 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.878136 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.879823 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.887808 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.896869 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.925315 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=23.925282627 podStartE2EDuration="23.925282627s" podCreationTimestamp="2026-03-20 09:02:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:02:59.925285547 +0000 UTC m=+346.611181961" watchObservedRunningTime="2026-03-20 09:02:59.925282627 +0000 UTC m=+346.611179031" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.971307 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.971801 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.998051 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-serving-cert\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.998169 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-client-ca\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.998211 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-proxy-ca-bundles\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.998251 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-client-ca\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.998393 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlb6r\" (UniqueName: \"kubernetes.io/projected/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-kube-api-access-hlb6r\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.998502 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-config\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.998575 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq49l\" (UniqueName: \"kubernetes.io/projected/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-kube-api-access-kq49l\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.998634 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-serving-cert\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:02:59 crc kubenswrapper[4823]: I0320 09:02:59.998894 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-config\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.015581 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.100305 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-serving-cert\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.100467 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-client-ca\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.100527 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-proxy-ca-bundles\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.100595 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-client-ca\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.100657 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlb6r\" (UniqueName: \"kubernetes.io/projected/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-kube-api-access-hlb6r\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.100712 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-config\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.100860 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq49l\" (UniqueName: \"kubernetes.io/projected/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-kube-api-access-kq49l\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.100925 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-serving-cert\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.101002 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-config\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.102310 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-config\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.102551 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-client-ca\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.102965 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-proxy-ca-bundles\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.103005 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-client-ca\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.103949 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-config\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.109385 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-serving-cert\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.115315 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-serving-cert\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.122409 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fca0d9b-55e0-49ae-a768-b2d1c19f1635" path="/var/lib/kubelet/pods/2fca0d9b-55e0-49ae-a768-b2d1c19f1635/volumes" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.124617 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66cb5542-0914-40bf-b27d-115a56b07012" path="/var/lib/kubelet/pods/66cb5542-0914-40bf-b27d-115a56b07012/volumes" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.129902 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq49l\" (UniqueName: \"kubernetes.io/projected/6043cc9c-56d2-4b40-9c07-74d0dbea2ff4-kube-api-access-kq49l\") pod \"route-controller-manager-7bb84fd79f-2845s\" (UID: \"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4\") " pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.130573 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddd258bb-2c2e-4211-94b3-947051b719f0" path="/var/lib/kubelet/pods/ddd258bb-2c2e-4211-94b3-947051b719f0/volumes" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.132277 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlb6r\" (UniqueName: \"kubernetes.io/projected/9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee-kube-api-access-hlb6r\") pod \"controller-manager-965c65cc7-q6vgf\" (UID: \"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee\") " pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.161756 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.206000 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.218276 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.291246 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.425184 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.449508 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.470031 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.581404 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.598872 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.613063 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.652912 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s"] Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.682761 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.700862 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-965c65cc7-q6vgf"] Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.752179 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.830714 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.843548 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" event={"ID":"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee","Type":"ContainerStarted","Data":"479aaba1d146b9ebfa34d16bffdb1fceccf6bb9443f494b276e21fa299afd6d0"} Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.843591 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" event={"ID":"9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee","Type":"ContainerStarted","Data":"5975c7c856675ef4c49ce0e7864d8915740273104366a6979739588179353694"} Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.843926 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.846528 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" event={"ID":"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4","Type":"ContainerStarted","Data":"ddeb448a78d46286acb7612348689e7e2ff967a3eb6f2b085fbd06e0a9ddf344"} Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.848472 4823 patch_prober.go:28] interesting pod/controller-manager-965c65cc7-q6vgf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.848584 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" podUID="9f5995f0-9abe-4e2a-9323-b1fe58a3b9ee" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Mar 20 09:03:00 crc kubenswrapper[4823]: I0320 09:03:00.862546 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" podStartSLOduration=43.86253075 podStartE2EDuration="43.86253075s" podCreationTimestamp="2026-03-20 09:02:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:03:00.86078548 +0000 UTC m=+347.546681894" watchObservedRunningTime="2026-03-20 09:03:00.86253075 +0000 UTC m=+347.548427124" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.107130 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.119552 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.483840 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.572565 4823 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.572674 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.572805 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.573800 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"19b0d3a42e7bec22ce3d90fee067806f5e1a58ddfd40e38ea8c13a134ff093e1"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.574024 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://19b0d3a42e7bec22ce3d90fee067806f5e1a58ddfd40e38ea8c13a134ff093e1" gracePeriod=30 Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.588376 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.589910 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.614915 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.748482 4823 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.792709 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.828791 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.854422 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" event={"ID":"6043cc9c-56d2-4b40-9c07-74d0dbea2ff4","Type":"ContainerStarted","Data":"c9d45cf63350b6c9951da7d9066e6711d26270fc3fe9a1df0b44d18adfb36403"} Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.859820 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-965c65cc7-q6vgf" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.877358 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" podStartSLOduration=44.877338339 podStartE2EDuration="44.877338339s" podCreationTimestamp="2026-03-20 09:02:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:03:01.877254077 +0000 UTC m=+348.563150481" watchObservedRunningTime="2026-03-20 09:03:01.877338339 +0000 UTC m=+348.563234713" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.890172 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 20 09:03:01 crc kubenswrapper[4823]: I0320 09:03:01.994872 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.050010 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-984c8fd85-9rw6v"] Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.051068 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.053547 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.053599 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.053553 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.053816 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.054972 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.054992 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.055041 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.055103 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.055130 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.057056 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.057356 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.057806 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.060299 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.074333 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.075304 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-984c8fd85-9rw6v"] Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.078828 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.131627 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-template-error\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.131694 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-template-login\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.131735 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-cliconfig\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.131803 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-audit-policies\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.131851 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.131886 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-session\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.131924 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-service-ca\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.131949 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.131975 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-audit-dir\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.132000 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-router-certs\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.132041 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.132110 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vv29\" (UniqueName: \"kubernetes.io/projected/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-kube-api-access-7vv29\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.132219 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.132274 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-serving-cert\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234082 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-router-certs\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234137 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234163 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vv29\" (UniqueName: \"kubernetes.io/projected/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-kube-api-access-7vv29\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234195 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234225 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-serving-cert\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234270 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-template-error\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234296 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-template-login\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234328 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-cliconfig\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234347 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-audit-policies\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234370 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234394 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-session\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234424 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-service-ca\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234445 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234495 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-audit-dir\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.234604 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-audit-dir\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.235598 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-cliconfig\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.235947 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-audit-policies\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.236171 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-service-ca\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.237832 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.238776 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.240680 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-template-login\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.240787 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-session\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.241592 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.241882 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-router-certs\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.242497 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-serving-cert\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.247118 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.247231 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-user-template-error\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.247332 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.252809 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vv29\" (UniqueName: \"kubernetes.io/projected/5e59e0ce-55cb-4d5f-9877-649b3bee6e77-kube-api-access-7vv29\") pod \"oauth-openshift-984c8fd85-9rw6v\" (UID: \"5e59e0ce-55cb-4d5f-9877-649b3bee6e77\") " pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.322935 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.390387 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.486928 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.496710 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.497227 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.574897 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.589944 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.676190 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.698433 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.778789 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.817833 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.842993 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-984c8fd85-9rw6v"] Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.851450 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.860189 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.865724 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7bb84fd79f-2845s" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.919149 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 20 09:03:02 crc kubenswrapper[4823]: I0320 09:03:02.986515 4823 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 20 09:03:03 crc kubenswrapper[4823]: I0320 09:03:03.180788 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 20 09:03:03 crc kubenswrapper[4823]: I0320 09:03:03.455551 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 20 09:03:03 crc kubenswrapper[4823]: I0320 09:03:03.868188 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" event={"ID":"5e59e0ce-55cb-4d5f-9877-649b3bee6e77","Type":"ContainerStarted","Data":"53a54f74f60665fe7a892d8b61261134a153b80b59925deefd0e5eb7a7e91250"} Mar 20 09:03:03 crc kubenswrapper[4823]: I0320 09:03:03.868274 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:03 crc kubenswrapper[4823]: I0320 09:03:03.868299 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" event={"ID":"5e59e0ce-55cb-4d5f-9877-649b3bee6e77","Type":"ContainerStarted","Data":"3d09d06ee0a1f4915cb67ba06143efd405ab20be0706e93508ddc1c857abcd53"} Mar 20 09:03:03 crc kubenswrapper[4823]: I0320 09:03:03.873892 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" Mar 20 09:03:03 crc kubenswrapper[4823]: I0320 09:03:03.897033 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-984c8fd85-9rw6v" podStartSLOduration=54.896999521 podStartE2EDuration="54.896999521s" podCreationTimestamp="2026-03-20 09:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:03:03.894260287 +0000 UTC m=+350.580156691" watchObservedRunningTime="2026-03-20 09:03:03.896999521 +0000 UTC m=+350.582895935" Mar 20 09:03:03 crc kubenswrapper[4823]: I0320 09:03:03.961101 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 20 09:03:04 crc kubenswrapper[4823]: I0320 09:03:04.170931 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 20 09:03:04 crc kubenswrapper[4823]: I0320 09:03:04.271359 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 20 09:03:04 crc kubenswrapper[4823]: I0320 09:03:04.277557 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 20 09:03:04 crc kubenswrapper[4823]: I0320 09:03:04.413734 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 20 09:03:04 crc kubenswrapper[4823]: I0320 09:03:04.435288 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 20 09:03:04 crc kubenswrapper[4823]: I0320 09:03:04.685049 4823 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 20 09:03:04 crc kubenswrapper[4823]: I0320 09:03:04.928461 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 20 09:03:04 crc kubenswrapper[4823]: I0320 09:03:04.948817 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 20 09:03:05 crc kubenswrapper[4823]: I0320 09:03:05.890403 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 20 09:03:05 crc kubenswrapper[4823]: I0320 09:03:05.894181 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 20 09:03:06 crc kubenswrapper[4823]: I0320 09:03:06.053162 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 20 09:03:10 crc kubenswrapper[4823]: I0320 09:03:10.238950 4823 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 20 09:03:10 crc kubenswrapper[4823]: I0320 09:03:10.239593 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490" gracePeriod=5 Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.849112 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.849912 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.945252 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.945332 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.945396 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.945450 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.945467 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.945520 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.945580 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.945586 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.945794 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.946109 4823 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.946147 4823 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.946173 4823 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.946198 4823 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.955289 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.955568 4823 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490" exitCode=137 Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.955628 4823 scope.go:117] "RemoveContainer" containerID="b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.955679 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:03:15 crc kubenswrapper[4823]: I0320 09:03:15.959964 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:03:16 crc kubenswrapper[4823]: I0320 09:03:16.006835 4823 scope.go:117] "RemoveContainer" containerID="b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490" Mar 20 09:03:16 crc kubenswrapper[4823]: E0320 09:03:16.007417 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490\": container with ID starting with b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490 not found: ID does not exist" containerID="b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490" Mar 20 09:03:16 crc kubenswrapper[4823]: I0320 09:03:16.007585 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490"} err="failed to get container status \"b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490\": rpc error: code = NotFound desc = could not find container \"b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490\": container with ID starting with b72af8287f76e93b42beaee9dc3b76fd9b887c3526cfe93b8dcef533800ad490 not found: ID does not exist" Mar 20 09:03:16 crc kubenswrapper[4823]: I0320 09:03:16.047417 4823 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:03:16 crc kubenswrapper[4823]: I0320 09:03:16.106828 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 20 09:03:16 crc kubenswrapper[4823]: I0320 09:03:16.107606 4823 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Mar 20 09:03:16 crc kubenswrapper[4823]: I0320 09:03:16.125016 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 20 09:03:16 crc kubenswrapper[4823]: I0320 09:03:16.125389 4823 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="48364e05-a3c9-4652-9f2e-b0edac550ece" Mar 20 09:03:16 crc kubenswrapper[4823]: I0320 09:03:16.130125 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 20 09:03:16 crc kubenswrapper[4823]: I0320 09:03:16.130161 4823 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="48364e05-a3c9-4652-9f2e-b0edac550ece" Mar 20 09:03:27 crc kubenswrapper[4823]: I0320 09:03:27.052002 4823 generic.go:334] "Generic (PLEG): container finished" podID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" containerID="7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67" exitCode=0 Mar 20 09:03:27 crc kubenswrapper[4823]: I0320 09:03:27.052125 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" event={"ID":"d31847cb-ea3f-46b1-ad2a-12d87b23d882","Type":"ContainerDied","Data":"7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67"} Mar 20 09:03:27 crc kubenswrapper[4823]: I0320 09:03:27.054671 4823 scope.go:117] "RemoveContainer" containerID="7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67" Mar 20 09:03:28 crc kubenswrapper[4823]: I0320 09:03:28.060295 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" event={"ID":"d31847cb-ea3f-46b1-ad2a-12d87b23d882","Type":"ContainerStarted","Data":"bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85"} Mar 20 09:03:28 crc kubenswrapper[4823]: I0320 09:03:28.060911 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:03:28 crc kubenswrapper[4823]: I0320 09:03:28.062030 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:03:32 crc kubenswrapper[4823]: I0320 09:03:32.090990 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 20 09:03:32 crc kubenswrapper[4823]: I0320 09:03:32.093173 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 09:03:32 crc kubenswrapper[4823]: I0320 09:03:32.093734 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 20 09:03:32 crc kubenswrapper[4823]: I0320 09:03:32.093790 4823 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="19b0d3a42e7bec22ce3d90fee067806f5e1a58ddfd40e38ea8c13a134ff093e1" exitCode=137 Mar 20 09:03:32 crc kubenswrapper[4823]: I0320 09:03:32.093824 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"19b0d3a42e7bec22ce3d90fee067806f5e1a58ddfd40e38ea8c13a134ff093e1"} Mar 20 09:03:32 crc kubenswrapper[4823]: I0320 09:03:32.093860 4823 scope.go:117] "RemoveContainer" containerID="2e64cdce0a2a1d3644da17bf462d0f58912611360f96fa8f3102684a61864483" Mar 20 09:03:33 crc kubenswrapper[4823]: I0320 09:03:33.103022 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 20 09:03:33 crc kubenswrapper[4823]: I0320 09:03:33.104795 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 09:03:33 crc kubenswrapper[4823]: I0320 09:03:33.104881 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"431ddd6ce956d17df5a77d3a084d0c5f533c8b5b7d6afd7081c758c4075a619a"} Mar 20 09:03:36 crc kubenswrapper[4823]: I0320 09:03:36.272843 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 09:03:41 crc kubenswrapper[4823]: I0320 09:03:41.572588 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 09:03:41 crc kubenswrapper[4823]: I0320 09:03:41.580587 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 09:03:42 crc kubenswrapper[4823]: I0320 09:03:42.169386 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.180718 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566624-7h65b"] Mar 20 09:04:00 crc kubenswrapper[4823]: E0320 09:04:00.182572 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.182677 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.182903 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.183487 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566624-7h65b" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.187569 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.187572 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.188462 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566624-7h65b"] Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.191433 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.289834 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trqsm\" (UniqueName: \"kubernetes.io/projected/2da31dce-84ef-4cc4-a959-97cbc515f685-kube-api-access-trqsm\") pod \"auto-csr-approver-29566624-7h65b\" (UID: \"2da31dce-84ef-4cc4-a959-97cbc515f685\") " pod="openshift-infra/auto-csr-approver-29566624-7h65b" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.391501 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trqsm\" (UniqueName: \"kubernetes.io/projected/2da31dce-84ef-4cc4-a959-97cbc515f685-kube-api-access-trqsm\") pod \"auto-csr-approver-29566624-7h65b\" (UID: \"2da31dce-84ef-4cc4-a959-97cbc515f685\") " pod="openshift-infra/auto-csr-approver-29566624-7h65b" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.410435 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trqsm\" (UniqueName: \"kubernetes.io/projected/2da31dce-84ef-4cc4-a959-97cbc515f685-kube-api-access-trqsm\") pod \"auto-csr-approver-29566624-7h65b\" (UID: \"2da31dce-84ef-4cc4-a959-97cbc515f685\") " pod="openshift-infra/auto-csr-approver-29566624-7h65b" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.509090 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566624-7h65b" Mar 20 09:04:00 crc kubenswrapper[4823]: I0320 09:04:00.974893 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566624-7h65b"] Mar 20 09:04:01 crc kubenswrapper[4823]: I0320 09:04:01.275900 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566624-7h65b" event={"ID":"2da31dce-84ef-4cc4-a959-97cbc515f685","Type":"ContainerStarted","Data":"30acf6c5671681116d62f44f26a4b7e057b3616d93135b7bd469e0728c763ae2"} Mar 20 09:04:02 crc kubenswrapper[4823]: I0320 09:04:02.284294 4823 generic.go:334] "Generic (PLEG): container finished" podID="2da31dce-84ef-4cc4-a959-97cbc515f685" containerID="efc76f0896d85f59ad395378397e44e69e669e1422379d4157219c62a9f1735f" exitCode=0 Mar 20 09:04:02 crc kubenswrapper[4823]: I0320 09:04:02.284468 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566624-7h65b" event={"ID":"2da31dce-84ef-4cc4-a959-97cbc515f685","Type":"ContainerDied","Data":"efc76f0896d85f59ad395378397e44e69e669e1422379d4157219c62a9f1735f"} Mar 20 09:04:03 crc kubenswrapper[4823]: I0320 09:04:03.621709 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566624-7h65b" Mar 20 09:04:03 crc kubenswrapper[4823]: I0320 09:04:03.732561 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trqsm\" (UniqueName: \"kubernetes.io/projected/2da31dce-84ef-4cc4-a959-97cbc515f685-kube-api-access-trqsm\") pod \"2da31dce-84ef-4cc4-a959-97cbc515f685\" (UID: \"2da31dce-84ef-4cc4-a959-97cbc515f685\") " Mar 20 09:04:03 crc kubenswrapper[4823]: I0320 09:04:03.737429 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2da31dce-84ef-4cc4-a959-97cbc515f685-kube-api-access-trqsm" (OuterVolumeSpecName: "kube-api-access-trqsm") pod "2da31dce-84ef-4cc4-a959-97cbc515f685" (UID: "2da31dce-84ef-4cc4-a959-97cbc515f685"). InnerVolumeSpecName "kube-api-access-trqsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:03 crc kubenswrapper[4823]: I0320 09:04:03.834245 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trqsm\" (UniqueName: \"kubernetes.io/projected/2da31dce-84ef-4cc4-a959-97cbc515f685-kube-api-access-trqsm\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:04 crc kubenswrapper[4823]: I0320 09:04:04.298035 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566624-7h65b" event={"ID":"2da31dce-84ef-4cc4-a959-97cbc515f685","Type":"ContainerDied","Data":"30acf6c5671681116d62f44f26a4b7e057b3616d93135b7bd469e0728c763ae2"} Mar 20 09:04:04 crc kubenswrapper[4823]: I0320 09:04:04.298108 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30acf6c5671681116d62f44f26a4b7e057b3616d93135b7bd469e0728c763ae2" Mar 20 09:04:04 crc kubenswrapper[4823]: I0320 09:04:04.298062 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566624-7h65b" Mar 20 09:04:09 crc kubenswrapper[4823]: I0320 09:04:09.430997 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:04:09 crc kubenswrapper[4823]: I0320 09:04:09.431279 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:04:39 crc kubenswrapper[4823]: I0320 09:04:39.430571 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:04:39 crc kubenswrapper[4823]: I0320 09:04:39.431584 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.519273 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7g599"] Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.520121 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7g599" podUID="95495063-7a29-4af1-a255-973a0129e289" containerName="registry-server" containerID="cri-o://f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf" gracePeriod=30 Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.526454 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k9bv8"] Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.526882 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k9bv8" podUID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" containerName="registry-server" containerID="cri-o://0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00" gracePeriod=30 Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.529551 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wd9j2"] Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.529851 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" podUID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" containerName="marketplace-operator" containerID="cri-o://bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85" gracePeriod=30 Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.539769 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6m5xl"] Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.540129 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6m5xl" podUID="2cee23f5-1268-4adf-8466-da5931de9825" containerName="registry-server" containerID="cri-o://9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a" gracePeriod=30 Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.541078 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vbgxd"] Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.543448 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vbgxd" podUID="694805b6-f8a1-4277-8132-100503fdc1a8" containerName="registry-server" containerID="cri-o://27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569" gracePeriod=30 Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.561400 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qcvh5"] Mar 20 09:04:55 crc kubenswrapper[4823]: E0320 09:04:55.561678 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2da31dce-84ef-4cc4-a959-97cbc515f685" containerName="oc" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.561694 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="2da31dce-84ef-4cc4-a959-97cbc515f685" containerName="oc" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.561840 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="2da31dce-84ef-4cc4-a959-97cbc515f685" containerName="oc" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.562305 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.580340 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qcvh5"] Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.709511 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mf79\" (UniqueName: \"kubernetes.io/projected/a5f8c83b-29f6-4dca-bee6-7bc332986168-kube-api-access-9mf79\") pod \"marketplace-operator-79b997595-qcvh5\" (UID: \"a5f8c83b-29f6-4dca-bee6-7bc332986168\") " pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.709573 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a5f8c83b-29f6-4dca-bee6-7bc332986168-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qcvh5\" (UID: \"a5f8c83b-29f6-4dca-bee6-7bc332986168\") " pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.709597 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a5f8c83b-29f6-4dca-bee6-7bc332986168-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qcvh5\" (UID: \"a5f8c83b-29f6-4dca-bee6-7bc332986168\") " pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.810845 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mf79\" (UniqueName: \"kubernetes.io/projected/a5f8c83b-29f6-4dca-bee6-7bc332986168-kube-api-access-9mf79\") pod \"marketplace-operator-79b997595-qcvh5\" (UID: \"a5f8c83b-29f6-4dca-bee6-7bc332986168\") " pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.810910 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a5f8c83b-29f6-4dca-bee6-7bc332986168-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qcvh5\" (UID: \"a5f8c83b-29f6-4dca-bee6-7bc332986168\") " pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.810932 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a5f8c83b-29f6-4dca-bee6-7bc332986168-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qcvh5\" (UID: \"a5f8c83b-29f6-4dca-bee6-7bc332986168\") " pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.812792 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a5f8c83b-29f6-4dca-bee6-7bc332986168-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qcvh5\" (UID: \"a5f8c83b-29f6-4dca-bee6-7bc332986168\") " pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.817460 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a5f8c83b-29f6-4dca-bee6-7bc332986168-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qcvh5\" (UID: \"a5f8c83b-29f6-4dca-bee6-7bc332986168\") " pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:55 crc kubenswrapper[4823]: I0320 09:04:55.828805 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mf79\" (UniqueName: \"kubernetes.io/projected/a5f8c83b-29f6-4dca-bee6-7bc332986168-kube-api-access-9mf79\") pod \"marketplace-operator-79b997595-qcvh5\" (UID: \"a5f8c83b-29f6-4dca-bee6-7bc332986168\") " pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.029590 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.051159 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.061635 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.062985 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.077884 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.093152 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214335 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l482s\" (UniqueName: \"kubernetes.io/projected/95495063-7a29-4af1-a255-973a0129e289-kube-api-access-l482s\") pod \"95495063-7a29-4af1-a255-973a0129e289\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214396 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtcp6\" (UniqueName: \"kubernetes.io/projected/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-kube-api-access-dtcp6\") pod \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214425 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-trusted-ca\") pod \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214469 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-operator-metrics\") pod \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214494 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-utilities\") pod \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214511 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzkf2\" (UniqueName: \"kubernetes.io/projected/2cee23f5-1268-4adf-8466-da5931de9825-kube-api-access-fzkf2\") pod \"2cee23f5-1268-4adf-8466-da5931de9825\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214547 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-catalog-content\") pod \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\" (UID: \"f9cc626b-cf2b-407f-ac25-c9ce01e0660d\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214574 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-utilities\") pod \"95495063-7a29-4af1-a255-973a0129e289\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214592 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-utilities\") pod \"2cee23f5-1268-4adf-8466-da5931de9825\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214617 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-catalog-content\") pod \"2cee23f5-1268-4adf-8466-da5931de9825\" (UID: \"2cee23f5-1268-4adf-8466-da5931de9825\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214631 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-utilities\") pod \"694805b6-f8a1-4277-8132-100503fdc1a8\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214648 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fxlf\" (UniqueName: \"kubernetes.io/projected/d31847cb-ea3f-46b1-ad2a-12d87b23d882-kube-api-access-6fxlf\") pod \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\" (UID: \"d31847cb-ea3f-46b1-ad2a-12d87b23d882\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214666 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc56b\" (UniqueName: \"kubernetes.io/projected/694805b6-f8a1-4277-8132-100503fdc1a8-kube-api-access-pc56b\") pod \"694805b6-f8a1-4277-8132-100503fdc1a8\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214691 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-catalog-content\") pod \"694805b6-f8a1-4277-8132-100503fdc1a8\" (UID: \"694805b6-f8a1-4277-8132-100503fdc1a8\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.214712 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-catalog-content\") pod \"95495063-7a29-4af1-a255-973a0129e289\" (UID: \"95495063-7a29-4af1-a255-973a0129e289\") " Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.215452 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "d31847cb-ea3f-46b1-ad2a-12d87b23d882" (UID: "d31847cb-ea3f-46b1-ad2a-12d87b23d882"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.215495 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-utilities" (OuterVolumeSpecName: "utilities") pod "f9cc626b-cf2b-407f-ac25-c9ce01e0660d" (UID: "f9cc626b-cf2b-407f-ac25-c9ce01e0660d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.215905 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-utilities" (OuterVolumeSpecName: "utilities") pod "2cee23f5-1268-4adf-8466-da5931de9825" (UID: "2cee23f5-1268-4adf-8466-da5931de9825"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.215918 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-utilities" (OuterVolumeSpecName: "utilities") pod "694805b6-f8a1-4277-8132-100503fdc1a8" (UID: "694805b6-f8a1-4277-8132-100503fdc1a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.217044 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-utilities" (OuterVolumeSpecName: "utilities") pod "95495063-7a29-4af1-a255-973a0129e289" (UID: "95495063-7a29-4af1-a255-973a0129e289"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.232872 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cee23f5-1268-4adf-8466-da5931de9825-kube-api-access-fzkf2" (OuterVolumeSpecName: "kube-api-access-fzkf2") pod "2cee23f5-1268-4adf-8466-da5931de9825" (UID: "2cee23f5-1268-4adf-8466-da5931de9825"). InnerVolumeSpecName "kube-api-access-fzkf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.233030 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-kube-api-access-dtcp6" (OuterVolumeSpecName: "kube-api-access-dtcp6") pod "f9cc626b-cf2b-407f-ac25-c9ce01e0660d" (UID: "f9cc626b-cf2b-407f-ac25-c9ce01e0660d"). InnerVolumeSpecName "kube-api-access-dtcp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.233084 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/694805b6-f8a1-4277-8132-100503fdc1a8-kube-api-access-pc56b" (OuterVolumeSpecName: "kube-api-access-pc56b") pod "694805b6-f8a1-4277-8132-100503fdc1a8" (UID: "694805b6-f8a1-4277-8132-100503fdc1a8"). InnerVolumeSpecName "kube-api-access-pc56b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.233226 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95495063-7a29-4af1-a255-973a0129e289-kube-api-access-l482s" (OuterVolumeSpecName: "kube-api-access-l482s") pod "95495063-7a29-4af1-a255-973a0129e289" (UID: "95495063-7a29-4af1-a255-973a0129e289"). InnerVolumeSpecName "kube-api-access-l482s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.233274 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "d31847cb-ea3f-46b1-ad2a-12d87b23d882" (UID: "d31847cb-ea3f-46b1-ad2a-12d87b23d882"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.233883 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d31847cb-ea3f-46b1-ad2a-12d87b23d882-kube-api-access-6fxlf" (OuterVolumeSpecName: "kube-api-access-6fxlf") pod "d31847cb-ea3f-46b1-ad2a-12d87b23d882" (UID: "d31847cb-ea3f-46b1-ad2a-12d87b23d882"). InnerVolumeSpecName "kube-api-access-6fxlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.268085 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2cee23f5-1268-4adf-8466-da5931de9825" (UID: "2cee23f5-1268-4adf-8466-da5931de9825"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.278717 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95495063-7a29-4af1-a255-973a0129e289" (UID: "95495063-7a29-4af1-a255-973a0129e289"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.294476 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qcvh5"] Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.300678 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9cc626b-cf2b-407f-ac25-c9ce01e0660d" (UID: "f9cc626b-cf2b-407f-ac25-c9ce01e0660d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316096 4823 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316127 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316138 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzkf2\" (UniqueName: \"kubernetes.io/projected/2cee23f5-1268-4adf-8466-da5931de9825-kube-api-access-fzkf2\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316146 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316154 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316162 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316170 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cee23f5-1268-4adf-8466-da5931de9825-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316192 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316203 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fxlf\" (UniqueName: \"kubernetes.io/projected/d31847cb-ea3f-46b1-ad2a-12d87b23d882-kube-api-access-6fxlf\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316215 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc56b\" (UniqueName: \"kubernetes.io/projected/694805b6-f8a1-4277-8132-100503fdc1a8-kube-api-access-pc56b\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316225 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95495063-7a29-4af1-a255-973a0129e289-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316236 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l482s\" (UniqueName: \"kubernetes.io/projected/95495063-7a29-4af1-a255-973a0129e289-kube-api-access-l482s\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316264 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtcp6\" (UniqueName: \"kubernetes.io/projected/f9cc626b-cf2b-407f-ac25-c9ce01e0660d-kube-api-access-dtcp6\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.316276 4823 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d31847cb-ea3f-46b1-ad2a-12d87b23d882-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.394703 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "694805b6-f8a1-4277-8132-100503fdc1a8" (UID: "694805b6-f8a1-4277-8132-100503fdc1a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.418039 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/694805b6-f8a1-4277-8132-100503fdc1a8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.632145 4823 generic.go:334] "Generic (PLEG): container finished" podID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" containerID="0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00" exitCode=0 Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.632224 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k9bv8" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.632243 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9bv8" event={"ID":"f9cc626b-cf2b-407f-ac25-c9ce01e0660d","Type":"ContainerDied","Data":"0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.632293 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k9bv8" event={"ID":"f9cc626b-cf2b-407f-ac25-c9ce01e0660d","Type":"ContainerDied","Data":"4d2487d8c4e1a7b21639e871db613c44418e15e647f5e5f4895aa1542523a8c0"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.632310 4823 scope.go:117] "RemoveContainer" containerID="0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.635369 4823 generic.go:334] "Generic (PLEG): container finished" podID="2cee23f5-1268-4adf-8466-da5931de9825" containerID="9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a" exitCode=0 Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.635434 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m5xl" event={"ID":"2cee23f5-1268-4adf-8466-da5931de9825","Type":"ContainerDied","Data":"9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.635440 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6m5xl" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.635452 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6m5xl" event={"ID":"2cee23f5-1268-4adf-8466-da5931de9825","Type":"ContainerDied","Data":"2fd3f02fd2f3d952dc91769865ba11a2ae3209c1459e4102154dd29a48b462f1"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.639274 4823 generic.go:334] "Generic (PLEG): container finished" podID="694805b6-f8a1-4277-8132-100503fdc1a8" containerID="27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569" exitCode=0 Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.639307 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbgxd" event={"ID":"694805b6-f8a1-4277-8132-100503fdc1a8","Type":"ContainerDied","Data":"27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.639321 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vbgxd" event={"ID":"694805b6-f8a1-4277-8132-100503fdc1a8","Type":"ContainerDied","Data":"b920167dcb12e251075f504088e9c1cd6db5443d9bf8f5370b4c75c86439d0db"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.639379 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vbgxd" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.645927 4823 generic.go:334] "Generic (PLEG): container finished" podID="95495063-7a29-4af1-a255-973a0129e289" containerID="f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf" exitCode=0 Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.645981 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g599" event={"ID":"95495063-7a29-4af1-a255-973a0129e289","Type":"ContainerDied","Data":"f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.646002 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g599" event={"ID":"95495063-7a29-4af1-a255-973a0129e289","Type":"ContainerDied","Data":"f6e1dbc892b868ff239a259c65c96139dd9b5a41762bae781e1fe154713c4cf2"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.646053 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7g599" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.654976 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" event={"ID":"a5f8c83b-29f6-4dca-bee6-7bc332986168","Type":"ContainerStarted","Data":"f7c4bddfbdab25b1202f55a120cacfac68f2478ede295833a3a8f6820e5ccd08"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.655257 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" event={"ID":"a5f8c83b-29f6-4dca-bee6-7bc332986168","Type":"ContainerStarted","Data":"79ff502e713045b4f1a02c19e5ae5b7015845ee2e6a686b413ad1418bbac4a65"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.655334 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.657075 4823 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qcvh5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.71:8080/healthz\": dial tcp 10.217.0.71:8080: connect: connection refused" start-of-body= Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.657117 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" podUID="a5f8c83b-29f6-4dca-bee6-7bc332986168" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.71:8080/healthz\": dial tcp 10.217.0.71:8080: connect: connection refused" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.659839 4823 generic.go:334] "Generic (PLEG): container finished" podID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" containerID="bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85" exitCode=0 Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.659877 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.660681 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" event={"ID":"d31847cb-ea3f-46b1-ad2a-12d87b23d882","Type":"ContainerDied","Data":"bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.660751 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wd9j2" event={"ID":"d31847cb-ea3f-46b1-ad2a-12d87b23d882","Type":"ContainerDied","Data":"3905787e3e265896fa9010989ce5bba4300bda58fb310d3a1c270f7c7eae46f2"} Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.680115 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" podStartSLOduration=1.680097081 podStartE2EDuration="1.680097081s" podCreationTimestamp="2026-03-20 09:04:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:04:56.677307891 +0000 UTC m=+463.363204275" watchObservedRunningTime="2026-03-20 09:04:56.680097081 +0000 UTC m=+463.365993455" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.682579 4823 scope.go:117] "RemoveContainer" containerID="e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.695107 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6m5xl"] Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.700471 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6m5xl"] Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.710693 4823 scope.go:117] "RemoveContainer" containerID="58ba417b79c9ab68df891927d11ee0bdf0405d4689ddeee8346c61122c0fd079" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.715468 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k9bv8"] Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.733759 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k9bv8"] Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.736179 4823 scope.go:117] "RemoveContainer" containerID="0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.736609 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00\": container with ID starting with 0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00 not found: ID does not exist" containerID="0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.736642 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00"} err="failed to get container status \"0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00\": rpc error: code = NotFound desc = could not find container \"0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00\": container with ID starting with 0f4a9389f861d060b7a3bf9446903d0a69f36ff88613f39a44423c89bf485f00 not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.736659 4823 scope.go:117] "RemoveContainer" containerID="e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.739417 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc\": container with ID starting with e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc not found: ID does not exist" containerID="e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.739466 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc"} err="failed to get container status \"e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc\": rpc error: code = NotFound desc = could not find container \"e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc\": container with ID starting with e12628b8108624e2f35947b501a885a75f42079f66e3101bdbce5e2251be94dc not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.739496 4823 scope.go:117] "RemoveContainer" containerID="58ba417b79c9ab68df891927d11ee0bdf0405d4689ddeee8346c61122c0fd079" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.740971 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wd9j2"] Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.741009 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58ba417b79c9ab68df891927d11ee0bdf0405d4689ddeee8346c61122c0fd079\": container with ID starting with 58ba417b79c9ab68df891927d11ee0bdf0405d4689ddeee8346c61122c0fd079 not found: ID does not exist" containerID="58ba417b79c9ab68df891927d11ee0bdf0405d4689ddeee8346c61122c0fd079" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.741167 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58ba417b79c9ab68df891927d11ee0bdf0405d4689ddeee8346c61122c0fd079"} err="failed to get container status \"58ba417b79c9ab68df891927d11ee0bdf0405d4689ddeee8346c61122c0fd079\": rpc error: code = NotFound desc = could not find container \"58ba417b79c9ab68df891927d11ee0bdf0405d4689ddeee8346c61122c0fd079\": container with ID starting with 58ba417b79c9ab68df891927d11ee0bdf0405d4689ddeee8346c61122c0fd079 not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.741199 4823 scope.go:117] "RemoveContainer" containerID="9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.754679 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wd9j2"] Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.758396 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vbgxd"] Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.761262 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vbgxd"] Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.764270 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7g599"] Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.764445 4823 scope.go:117] "RemoveContainer" containerID="af7f3dcbbfcda86f18af6a5e208c415b74b95b55440662f61f73fe293e2cd466" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.768822 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7g599"] Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.790588 4823 scope.go:117] "RemoveContainer" containerID="b9674954a944094d62dfc8be3a4182df4ae1cafc9d7ee9c762ad9a7baacd9f3d" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.812448 4823 scope.go:117] "RemoveContainer" containerID="9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.812891 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a\": container with ID starting with 9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a not found: ID does not exist" containerID="9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.812927 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a"} err="failed to get container status \"9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a\": rpc error: code = NotFound desc = could not find container \"9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a\": container with ID starting with 9359a50ee421dedc470c3b4204323b44ff980c9807df3d485fbd99a961e9440a not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.812959 4823 scope.go:117] "RemoveContainer" containerID="af7f3dcbbfcda86f18af6a5e208c415b74b95b55440662f61f73fe293e2cd466" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.813225 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af7f3dcbbfcda86f18af6a5e208c415b74b95b55440662f61f73fe293e2cd466\": container with ID starting with af7f3dcbbfcda86f18af6a5e208c415b74b95b55440662f61f73fe293e2cd466 not found: ID does not exist" containerID="af7f3dcbbfcda86f18af6a5e208c415b74b95b55440662f61f73fe293e2cd466" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.813276 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af7f3dcbbfcda86f18af6a5e208c415b74b95b55440662f61f73fe293e2cd466"} err="failed to get container status \"af7f3dcbbfcda86f18af6a5e208c415b74b95b55440662f61f73fe293e2cd466\": rpc error: code = NotFound desc = could not find container \"af7f3dcbbfcda86f18af6a5e208c415b74b95b55440662f61f73fe293e2cd466\": container with ID starting with af7f3dcbbfcda86f18af6a5e208c415b74b95b55440662f61f73fe293e2cd466 not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.813312 4823 scope.go:117] "RemoveContainer" containerID="b9674954a944094d62dfc8be3a4182df4ae1cafc9d7ee9c762ad9a7baacd9f3d" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.813550 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9674954a944094d62dfc8be3a4182df4ae1cafc9d7ee9c762ad9a7baacd9f3d\": container with ID starting with b9674954a944094d62dfc8be3a4182df4ae1cafc9d7ee9c762ad9a7baacd9f3d not found: ID does not exist" containerID="b9674954a944094d62dfc8be3a4182df4ae1cafc9d7ee9c762ad9a7baacd9f3d" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.813580 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9674954a944094d62dfc8be3a4182df4ae1cafc9d7ee9c762ad9a7baacd9f3d"} err="failed to get container status \"b9674954a944094d62dfc8be3a4182df4ae1cafc9d7ee9c762ad9a7baacd9f3d\": rpc error: code = NotFound desc = could not find container \"b9674954a944094d62dfc8be3a4182df4ae1cafc9d7ee9c762ad9a7baacd9f3d\": container with ID starting with b9674954a944094d62dfc8be3a4182df4ae1cafc9d7ee9c762ad9a7baacd9f3d not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.813600 4823 scope.go:117] "RemoveContainer" containerID="27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.830920 4823 scope.go:117] "RemoveContainer" containerID="a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.848528 4823 scope.go:117] "RemoveContainer" containerID="fa834b17f4012d635f99377eb4f8ace23e290542189ef037ee00b85fe968ab1f" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.860422 4823 scope.go:117] "RemoveContainer" containerID="27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.860769 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569\": container with ID starting with 27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569 not found: ID does not exist" containerID="27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.860809 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569"} err="failed to get container status \"27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569\": rpc error: code = NotFound desc = could not find container \"27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569\": container with ID starting with 27042fb824de13fae5b79a1b762137ea84084042a60fdff2a17230431f582569 not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.860835 4823 scope.go:117] "RemoveContainer" containerID="a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.861266 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5\": container with ID starting with a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5 not found: ID does not exist" containerID="a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.861304 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5"} err="failed to get container status \"a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5\": rpc error: code = NotFound desc = could not find container \"a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5\": container with ID starting with a0ebdd17cd0de5bb6d41c9120b6e0ac660647d0ba3755b6e534376576ffe34f5 not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.861348 4823 scope.go:117] "RemoveContainer" containerID="fa834b17f4012d635f99377eb4f8ace23e290542189ef037ee00b85fe968ab1f" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.861732 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa834b17f4012d635f99377eb4f8ace23e290542189ef037ee00b85fe968ab1f\": container with ID starting with fa834b17f4012d635f99377eb4f8ace23e290542189ef037ee00b85fe968ab1f not found: ID does not exist" containerID="fa834b17f4012d635f99377eb4f8ace23e290542189ef037ee00b85fe968ab1f" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.861788 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa834b17f4012d635f99377eb4f8ace23e290542189ef037ee00b85fe968ab1f"} err="failed to get container status \"fa834b17f4012d635f99377eb4f8ace23e290542189ef037ee00b85fe968ab1f\": rpc error: code = NotFound desc = could not find container \"fa834b17f4012d635f99377eb4f8ace23e290542189ef037ee00b85fe968ab1f\": container with ID starting with fa834b17f4012d635f99377eb4f8ace23e290542189ef037ee00b85fe968ab1f not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.861814 4823 scope.go:117] "RemoveContainer" containerID="f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.874701 4823 scope.go:117] "RemoveContainer" containerID="cc290a0fed4218f661268d816343115152c92147297093a88527cf98238659eb" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.887068 4823 scope.go:117] "RemoveContainer" containerID="a54dd9c64820ad8e9843816ed6926c74b7e3d824b002b853cd38c2200d1f9bbc" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.898399 4823 scope.go:117] "RemoveContainer" containerID="f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.898840 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf\": container with ID starting with f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf not found: ID does not exist" containerID="f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.898877 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf"} err="failed to get container status \"f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf\": rpc error: code = NotFound desc = could not find container \"f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf\": container with ID starting with f89eb56413c525b2638e4d6a05012ed27e8dc47a0a6b2d6cdb93b53dc332d2bf not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.898905 4823 scope.go:117] "RemoveContainer" containerID="cc290a0fed4218f661268d816343115152c92147297093a88527cf98238659eb" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.899257 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc290a0fed4218f661268d816343115152c92147297093a88527cf98238659eb\": container with ID starting with cc290a0fed4218f661268d816343115152c92147297093a88527cf98238659eb not found: ID does not exist" containerID="cc290a0fed4218f661268d816343115152c92147297093a88527cf98238659eb" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.899299 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc290a0fed4218f661268d816343115152c92147297093a88527cf98238659eb"} err="failed to get container status \"cc290a0fed4218f661268d816343115152c92147297093a88527cf98238659eb\": rpc error: code = NotFound desc = could not find container \"cc290a0fed4218f661268d816343115152c92147297093a88527cf98238659eb\": container with ID starting with cc290a0fed4218f661268d816343115152c92147297093a88527cf98238659eb not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.899329 4823 scope.go:117] "RemoveContainer" containerID="a54dd9c64820ad8e9843816ed6926c74b7e3d824b002b853cd38c2200d1f9bbc" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.899578 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a54dd9c64820ad8e9843816ed6926c74b7e3d824b002b853cd38c2200d1f9bbc\": container with ID starting with a54dd9c64820ad8e9843816ed6926c74b7e3d824b002b853cd38c2200d1f9bbc not found: ID does not exist" containerID="a54dd9c64820ad8e9843816ed6926c74b7e3d824b002b853cd38c2200d1f9bbc" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.899625 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54dd9c64820ad8e9843816ed6926c74b7e3d824b002b853cd38c2200d1f9bbc"} err="failed to get container status \"a54dd9c64820ad8e9843816ed6926c74b7e3d824b002b853cd38c2200d1f9bbc\": rpc error: code = NotFound desc = could not find container \"a54dd9c64820ad8e9843816ed6926c74b7e3d824b002b853cd38c2200d1f9bbc\": container with ID starting with a54dd9c64820ad8e9843816ed6926c74b7e3d824b002b853cd38c2200d1f9bbc not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.899641 4823 scope.go:117] "RemoveContainer" containerID="bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.917064 4823 scope.go:117] "RemoveContainer" containerID="7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.929311 4823 scope.go:117] "RemoveContainer" containerID="bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.932365 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85\": container with ID starting with bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85 not found: ID does not exist" containerID="bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.932407 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85"} err="failed to get container status \"bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85\": rpc error: code = NotFound desc = could not find container \"bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85\": container with ID starting with bb4541af2496d8b8075d3d1ec076c657ba55ec55719d3ef53251336b7ff7db85 not found: ID does not exist" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.932433 4823 scope.go:117] "RemoveContainer" containerID="7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67" Mar 20 09:04:56 crc kubenswrapper[4823]: E0320 09:04:56.932807 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67\": container with ID starting with 7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67 not found: ID does not exist" containerID="7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67" Mar 20 09:04:56 crc kubenswrapper[4823]: I0320 09:04:56.932862 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67"} err="failed to get container status \"7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67\": rpc error: code = NotFound desc = could not find container \"7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67\": container with ID starting with 7130b8835fdd6755e52bb8205e35ffa4e41fcec5831f6992a1e66795f8d45a67 not found: ID does not exist" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.679591 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qcvh5" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729000 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j8d5w"] Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729265 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="694805b6-f8a1-4277-8132-100503fdc1a8" containerName="extract-utilities" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729289 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="694805b6-f8a1-4277-8132-100503fdc1a8" containerName="extract-utilities" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729309 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="694805b6-f8a1-4277-8132-100503fdc1a8" containerName="extract-content" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729316 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="694805b6-f8a1-4277-8132-100503fdc1a8" containerName="extract-content" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729327 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cee23f5-1268-4adf-8466-da5931de9825" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729334 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cee23f5-1268-4adf-8466-da5931de9825" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729346 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" containerName="marketplace-operator" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729354 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" containerName="marketplace-operator" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729363 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="694805b6-f8a1-4277-8132-100503fdc1a8" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729373 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="694805b6-f8a1-4277-8132-100503fdc1a8" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729382 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729389 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729400 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" containerName="marketplace-operator" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729407 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" containerName="marketplace-operator" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729419 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95495063-7a29-4af1-a255-973a0129e289" containerName="extract-utilities" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729426 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="95495063-7a29-4af1-a255-973a0129e289" containerName="extract-utilities" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729437 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95495063-7a29-4af1-a255-973a0129e289" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729444 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="95495063-7a29-4af1-a255-973a0129e289" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729455 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cee23f5-1268-4adf-8466-da5931de9825" containerName="extract-content" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729462 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cee23f5-1268-4adf-8466-da5931de9825" containerName="extract-content" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729475 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cee23f5-1268-4adf-8466-da5931de9825" containerName="extract-utilities" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729482 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cee23f5-1268-4adf-8466-da5931de9825" containerName="extract-utilities" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729490 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" containerName="extract-content" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729498 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" containerName="extract-content" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729509 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95495063-7a29-4af1-a255-973a0129e289" containerName="extract-content" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729517 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="95495063-7a29-4af1-a255-973a0129e289" containerName="extract-content" Mar 20 09:04:57 crc kubenswrapper[4823]: E0320 09:04:57.729530 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" containerName="extract-utilities" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729538 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" containerName="extract-utilities" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729638 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729650 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" containerName="marketplace-operator" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729659 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="95495063-7a29-4af1-a255-973a0129e289" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729674 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" containerName="marketplace-operator" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729686 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="694805b6-f8a1-4277-8132-100503fdc1a8" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.729696 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cee23f5-1268-4adf-8466-da5931de9825" containerName="registry-server" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.730579 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.732551 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.742648 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j8d5w"] Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.841838 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acd3ec8c-1372-46ef-9c71-9c6185bfd892-catalog-content\") pod \"redhat-marketplace-j8d5w\" (UID: \"acd3ec8c-1372-46ef-9c71-9c6185bfd892\") " pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.841882 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8czb\" (UniqueName: \"kubernetes.io/projected/acd3ec8c-1372-46ef-9c71-9c6185bfd892-kube-api-access-r8czb\") pod \"redhat-marketplace-j8d5w\" (UID: \"acd3ec8c-1372-46ef-9c71-9c6185bfd892\") " pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.841929 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acd3ec8c-1372-46ef-9c71-9c6185bfd892-utilities\") pod \"redhat-marketplace-j8d5w\" (UID: \"acd3ec8c-1372-46ef-9c71-9c6185bfd892\") " pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.923016 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wzt57"] Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.923905 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.926267 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.937940 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wzt57"] Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.942837 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acd3ec8c-1372-46ef-9c71-9c6185bfd892-utilities\") pod \"redhat-marketplace-j8d5w\" (UID: \"acd3ec8c-1372-46ef-9c71-9c6185bfd892\") " pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.942987 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acd3ec8c-1372-46ef-9c71-9c6185bfd892-catalog-content\") pod \"redhat-marketplace-j8d5w\" (UID: \"acd3ec8c-1372-46ef-9c71-9c6185bfd892\") " pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.943019 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8czb\" (UniqueName: \"kubernetes.io/projected/acd3ec8c-1372-46ef-9c71-9c6185bfd892-kube-api-access-r8czb\") pod \"redhat-marketplace-j8d5w\" (UID: \"acd3ec8c-1372-46ef-9c71-9c6185bfd892\") " pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.943472 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acd3ec8c-1372-46ef-9c71-9c6185bfd892-utilities\") pod \"redhat-marketplace-j8d5w\" (UID: \"acd3ec8c-1372-46ef-9c71-9c6185bfd892\") " pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.943573 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acd3ec8c-1372-46ef-9c71-9c6185bfd892-catalog-content\") pod \"redhat-marketplace-j8d5w\" (UID: \"acd3ec8c-1372-46ef-9c71-9c6185bfd892\") " pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:04:57 crc kubenswrapper[4823]: I0320 09:04:57.965195 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8czb\" (UniqueName: \"kubernetes.io/projected/acd3ec8c-1372-46ef-9c71-9c6185bfd892-kube-api-access-r8czb\") pod \"redhat-marketplace-j8d5w\" (UID: \"acd3ec8c-1372-46ef-9c71-9c6185bfd892\") " pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.044387 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfwdv\" (UniqueName: \"kubernetes.io/projected/0a6b6c50-171e-4eec-acf0-29dc3406e657-kube-api-access-tfwdv\") pod \"redhat-operators-wzt57\" (UID: \"0a6b6c50-171e-4eec-acf0-29dc3406e657\") " pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.044595 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a6b6c50-171e-4eec-acf0-29dc3406e657-catalog-content\") pod \"redhat-operators-wzt57\" (UID: \"0a6b6c50-171e-4eec-acf0-29dc3406e657\") " pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.044687 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a6b6c50-171e-4eec-acf0-29dc3406e657-utilities\") pod \"redhat-operators-wzt57\" (UID: \"0a6b6c50-171e-4eec-acf0-29dc3406e657\") " pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.056676 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.107531 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cee23f5-1268-4adf-8466-da5931de9825" path="/var/lib/kubelet/pods/2cee23f5-1268-4adf-8466-da5931de9825/volumes" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.108948 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="694805b6-f8a1-4277-8132-100503fdc1a8" path="/var/lib/kubelet/pods/694805b6-f8a1-4277-8132-100503fdc1a8/volumes" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.110088 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95495063-7a29-4af1-a255-973a0129e289" path="/var/lib/kubelet/pods/95495063-7a29-4af1-a255-973a0129e289/volumes" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.112398 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d31847cb-ea3f-46b1-ad2a-12d87b23d882" path="/var/lib/kubelet/pods/d31847cb-ea3f-46b1-ad2a-12d87b23d882/volumes" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.113345 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9cc626b-cf2b-407f-ac25-c9ce01e0660d" path="/var/lib/kubelet/pods/f9cc626b-cf2b-407f-ac25-c9ce01e0660d/volumes" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.145838 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a6b6c50-171e-4eec-acf0-29dc3406e657-catalog-content\") pod \"redhat-operators-wzt57\" (UID: \"0a6b6c50-171e-4eec-acf0-29dc3406e657\") " pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.146272 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a6b6c50-171e-4eec-acf0-29dc3406e657-utilities\") pod \"redhat-operators-wzt57\" (UID: \"0a6b6c50-171e-4eec-acf0-29dc3406e657\") " pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.146337 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a6b6c50-171e-4eec-acf0-29dc3406e657-catalog-content\") pod \"redhat-operators-wzt57\" (UID: \"0a6b6c50-171e-4eec-acf0-29dc3406e657\") " pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.146394 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfwdv\" (UniqueName: \"kubernetes.io/projected/0a6b6c50-171e-4eec-acf0-29dc3406e657-kube-api-access-tfwdv\") pod \"redhat-operators-wzt57\" (UID: \"0a6b6c50-171e-4eec-acf0-29dc3406e657\") " pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.146719 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a6b6c50-171e-4eec-acf0-29dc3406e657-utilities\") pod \"redhat-operators-wzt57\" (UID: \"0a6b6c50-171e-4eec-acf0-29dc3406e657\") " pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.181357 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfwdv\" (UniqueName: \"kubernetes.io/projected/0a6b6c50-171e-4eec-acf0-29dc3406e657-kube-api-access-tfwdv\") pod \"redhat-operators-wzt57\" (UID: \"0a6b6c50-171e-4eec-acf0-29dc3406e657\") " pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.245158 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.276975 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j8d5w"] Mar 20 09:04:58 crc kubenswrapper[4823]: W0320 09:04:58.284399 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacd3ec8c_1372_46ef_9c71_9c6185bfd892.slice/crio-b63c211bf10203e6439f21629f35742449af4a5b6c1b5c9291d3ddabdd085800 WatchSource:0}: Error finding container b63c211bf10203e6439f21629f35742449af4a5b6c1b5c9291d3ddabdd085800: Status 404 returned error can't find the container with id b63c211bf10203e6439f21629f35742449af4a5b6c1b5c9291d3ddabdd085800 Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.478118 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wzt57"] Mar 20 09:04:58 crc kubenswrapper[4823]: W0320 09:04:58.523229 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a6b6c50_171e_4eec_acf0_29dc3406e657.slice/crio-983c2ab952a6459683e1ea1a41d9cddcbdd4308aaaaf640c74611de8a00a6d90 WatchSource:0}: Error finding container 983c2ab952a6459683e1ea1a41d9cddcbdd4308aaaaf640c74611de8a00a6d90: Status 404 returned error can't find the container with id 983c2ab952a6459683e1ea1a41d9cddcbdd4308aaaaf640c74611de8a00a6d90 Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.681792 4823 generic.go:334] "Generic (PLEG): container finished" podID="0a6b6c50-171e-4eec-acf0-29dc3406e657" containerID="9b8d037c5da4b788c02a5ff374742b5de6b1a964aad040472608068423c14b0a" exitCode=0 Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.682022 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzt57" event={"ID":"0a6b6c50-171e-4eec-acf0-29dc3406e657","Type":"ContainerDied","Data":"9b8d037c5da4b788c02a5ff374742b5de6b1a964aad040472608068423c14b0a"} Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.686100 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzt57" event={"ID":"0a6b6c50-171e-4eec-acf0-29dc3406e657","Type":"ContainerStarted","Data":"983c2ab952a6459683e1ea1a41d9cddcbdd4308aaaaf640c74611de8a00a6d90"} Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.686259 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j8d5w" event={"ID":"acd3ec8c-1372-46ef-9c71-9c6185bfd892","Type":"ContainerDied","Data":"605024b2c54498c82082627d737f44f03e2394ddac0b9f136c01756956355737"} Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.686005 4823 generic.go:334] "Generic (PLEG): container finished" podID="acd3ec8c-1372-46ef-9c71-9c6185bfd892" containerID="605024b2c54498c82082627d737f44f03e2394ddac0b9f136c01756956355737" exitCode=0 Mar 20 09:04:58 crc kubenswrapper[4823]: I0320 09:04:58.687860 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j8d5w" event={"ID":"acd3ec8c-1372-46ef-9c71-9c6185bfd892","Type":"ContainerStarted","Data":"b63c211bf10203e6439f21629f35742449af4a5b6c1b5c9291d3ddabdd085800"} Mar 20 09:04:59 crc kubenswrapper[4823]: I0320 09:04:59.696599 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzt57" event={"ID":"0a6b6c50-171e-4eec-acf0-29dc3406e657","Type":"ContainerStarted","Data":"e58de630236018e6e0fc56f06328a140db63753fa01829b8d3d9337035d87868"} Mar 20 09:04:59 crc kubenswrapper[4823]: I0320 09:04:59.699725 4823 generic.go:334] "Generic (PLEG): container finished" podID="acd3ec8c-1372-46ef-9c71-9c6185bfd892" containerID="99da7e47a6fb78881203ef333954ebb9679af5751f419e852728deeac0d49be8" exitCode=0 Mar 20 09:04:59 crc kubenswrapper[4823]: I0320 09:04:59.699784 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j8d5w" event={"ID":"acd3ec8c-1372-46ef-9c71-9c6185bfd892","Type":"ContainerDied","Data":"99da7e47a6fb78881203ef333954ebb9679af5751f419e852728deeac0d49be8"} Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.133518 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f5t7w"] Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.134486 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.136448 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.147270 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5t7w"] Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.276372 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-772sl\" (UniqueName: \"kubernetes.io/projected/f5c7111d-329b-45f4-bbd8-643fe3533f79-kube-api-access-772sl\") pod \"community-operators-f5t7w\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.276478 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-catalog-content\") pod \"community-operators-f5t7w\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.276560 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-utilities\") pod \"community-operators-f5t7w\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.331298 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2gqwq"] Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.332311 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.334357 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.377706 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-772sl\" (UniqueName: \"kubernetes.io/projected/f5c7111d-329b-45f4-bbd8-643fe3533f79-kube-api-access-772sl\") pod \"community-operators-f5t7w\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.377782 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-catalog-content\") pod \"community-operators-f5t7w\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.377820 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-utilities\") pod \"community-operators-f5t7w\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.378267 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-utilities\") pod \"community-operators-f5t7w\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.378839 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-catalog-content\") pod \"community-operators-f5t7w\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.383529 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2gqwq"] Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.409512 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-772sl\" (UniqueName: \"kubernetes.io/projected/f5c7111d-329b-45f4-bbd8-643fe3533f79-kube-api-access-772sl\") pod \"community-operators-f5t7w\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.452263 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.479605 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162-utilities\") pod \"certified-operators-2gqwq\" (UID: \"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162\") " pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.479680 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxw29\" (UniqueName: \"kubernetes.io/projected/1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162-kube-api-access-jxw29\") pod \"certified-operators-2gqwq\" (UID: \"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162\") " pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.479793 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162-catalog-content\") pod \"certified-operators-2gqwq\" (UID: \"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162\") " pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.580987 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162-catalog-content\") pod \"certified-operators-2gqwq\" (UID: \"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162\") " pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.581170 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162-utilities\") pod \"certified-operators-2gqwq\" (UID: \"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162\") " pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.581242 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxw29\" (UniqueName: \"kubernetes.io/projected/1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162-kube-api-access-jxw29\") pod \"certified-operators-2gqwq\" (UID: \"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162\") " pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.582644 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162-catalog-content\") pod \"certified-operators-2gqwq\" (UID: \"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162\") " pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.583128 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162-utilities\") pod \"certified-operators-2gqwq\" (UID: \"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162\") " pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.598134 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxw29\" (UniqueName: \"kubernetes.io/projected/1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162-kube-api-access-jxw29\") pod \"certified-operators-2gqwq\" (UID: \"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162\") " pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.691775 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.714644 4823 generic.go:334] "Generic (PLEG): container finished" podID="0a6b6c50-171e-4eec-acf0-29dc3406e657" containerID="e58de630236018e6e0fc56f06328a140db63753fa01829b8d3d9337035d87868" exitCode=0 Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.714707 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzt57" event={"ID":"0a6b6c50-171e-4eec-acf0-29dc3406e657","Type":"ContainerDied","Data":"e58de630236018e6e0fc56f06328a140db63753fa01829b8d3d9337035d87868"} Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.719613 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j8d5w" event={"ID":"acd3ec8c-1372-46ef-9c71-9c6185bfd892","Type":"ContainerStarted","Data":"d30a45df735969c4d76af9f75b6ed9e0068910a375e9729671cdf4c7e6143c16"} Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.756417 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j8d5w" podStartSLOduration=2.281655048 podStartE2EDuration="3.75639152s" podCreationTimestamp="2026-03-20 09:04:57 +0000 UTC" firstStartedPulling="2026-03-20 09:04:58.687204267 +0000 UTC m=+465.373100651" lastFinishedPulling="2026-03-20 09:05:00.161940749 +0000 UTC m=+466.847837123" observedRunningTime="2026-03-20 09:05:00.755112739 +0000 UTC m=+467.441009143" watchObservedRunningTime="2026-03-20 09:05:00.75639152 +0000 UTC m=+467.442287904" Mar 20 09:05:00 crc kubenswrapper[4823]: W0320 09:05:00.838176 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5c7111d_329b_45f4_bbd8_643fe3533f79.slice/crio-03192cd0d2e36348251c0152afe01a40bfdbb91098c60b5bebf9e42954eac82f WatchSource:0}: Error finding container 03192cd0d2e36348251c0152afe01a40bfdbb91098c60b5bebf9e42954eac82f: Status 404 returned error can't find the container with id 03192cd0d2e36348251c0152afe01a40bfdbb91098c60b5bebf9e42954eac82f Mar 20 09:05:00 crc kubenswrapper[4823]: I0320 09:05:00.840372 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5t7w"] Mar 20 09:05:01 crc kubenswrapper[4823]: I0320 09:05:01.105556 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2gqwq"] Mar 20 09:05:01 crc kubenswrapper[4823]: I0320 09:05:01.726059 4823 generic.go:334] "Generic (PLEG): container finished" podID="1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162" containerID="ca7e53aa7ed90a7844da3ddacdcc62ee091e2b93cda3c46815176865e3844a4a" exitCode=0 Mar 20 09:05:01 crc kubenswrapper[4823]: I0320 09:05:01.726156 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2gqwq" event={"ID":"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162","Type":"ContainerDied","Data":"ca7e53aa7ed90a7844da3ddacdcc62ee091e2b93cda3c46815176865e3844a4a"} Mar 20 09:05:01 crc kubenswrapper[4823]: I0320 09:05:01.726475 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2gqwq" event={"ID":"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162","Type":"ContainerStarted","Data":"5411a62d3230f2b6133a16ccad99a49a501edda3ce0f4a21e004010386819437"} Mar 20 09:05:01 crc kubenswrapper[4823]: I0320 09:05:01.732828 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wzt57" event={"ID":"0a6b6c50-171e-4eec-acf0-29dc3406e657","Type":"ContainerStarted","Data":"98f80cfd8d709887105693a1b969756d765503d6de20aa439c31e27761fa658f"} Mar 20 09:05:01 crc kubenswrapper[4823]: I0320 09:05:01.734626 4823 generic.go:334] "Generic (PLEG): container finished" podID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerID="b71bc856792b5242e62166acea56c00d0f4165624bca73121fda8b33d1ad94e3" exitCode=0 Mar 20 09:05:01 crc kubenswrapper[4823]: I0320 09:05:01.734666 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5t7w" event={"ID":"f5c7111d-329b-45f4-bbd8-643fe3533f79","Type":"ContainerDied","Data":"b71bc856792b5242e62166acea56c00d0f4165624bca73121fda8b33d1ad94e3"} Mar 20 09:05:01 crc kubenswrapper[4823]: I0320 09:05:01.734700 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5t7w" event={"ID":"f5c7111d-329b-45f4-bbd8-643fe3533f79","Type":"ContainerStarted","Data":"03192cd0d2e36348251c0152afe01a40bfdbb91098c60b5bebf9e42954eac82f"} Mar 20 09:05:01 crc kubenswrapper[4823]: I0320 09:05:01.800160 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wzt57" podStartSLOduration=2.169522348 podStartE2EDuration="4.800137724s" podCreationTimestamp="2026-03-20 09:04:57 +0000 UTC" firstStartedPulling="2026-03-20 09:04:58.684245712 +0000 UTC m=+465.370142086" lastFinishedPulling="2026-03-20 09:05:01.314861088 +0000 UTC m=+468.000757462" observedRunningTime="2026-03-20 09:05:01.787847711 +0000 UTC m=+468.473744095" watchObservedRunningTime="2026-03-20 09:05:01.800137724 +0000 UTC m=+468.486034108" Mar 20 09:05:02 crc kubenswrapper[4823]: I0320 09:05:02.746926 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5t7w" event={"ID":"f5c7111d-329b-45f4-bbd8-643fe3533f79","Type":"ContainerStarted","Data":"907f33c754629028b717dbbcead8f55dfe6305750aa27e1ca3804b4ce5b039a1"} Mar 20 09:05:03 crc kubenswrapper[4823]: I0320 09:05:03.754679 4823 generic.go:334] "Generic (PLEG): container finished" podID="1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162" containerID="ca70bfcb8a14170203ab4c384b4c536534674cc5f08fbe35772efc3ad96c91c3" exitCode=0 Mar 20 09:05:03 crc kubenswrapper[4823]: I0320 09:05:03.754801 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2gqwq" event={"ID":"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162","Type":"ContainerDied","Data":"ca70bfcb8a14170203ab4c384b4c536534674cc5f08fbe35772efc3ad96c91c3"} Mar 20 09:05:03 crc kubenswrapper[4823]: I0320 09:05:03.759073 4823 generic.go:334] "Generic (PLEG): container finished" podID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerID="907f33c754629028b717dbbcead8f55dfe6305750aa27e1ca3804b4ce5b039a1" exitCode=0 Mar 20 09:05:03 crc kubenswrapper[4823]: I0320 09:05:03.759120 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5t7w" event={"ID":"f5c7111d-329b-45f4-bbd8-643fe3533f79","Type":"ContainerDied","Data":"907f33c754629028b717dbbcead8f55dfe6305750aa27e1ca3804b4ce5b039a1"} Mar 20 09:05:04 crc kubenswrapper[4823]: I0320 09:05:04.767402 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5t7w" event={"ID":"f5c7111d-329b-45f4-bbd8-643fe3533f79","Type":"ContainerStarted","Data":"23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc"} Mar 20 09:05:04 crc kubenswrapper[4823]: I0320 09:05:04.770522 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2gqwq" event={"ID":"1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162","Type":"ContainerStarted","Data":"8d71042b9d939bdfc6a11665aca397890cd45ca48743fcc985b0f19664f29d95"} Mar 20 09:05:04 crc kubenswrapper[4823]: I0320 09:05:04.802952 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f5t7w" podStartSLOduration=2.397766487 podStartE2EDuration="4.802931256s" podCreationTimestamp="2026-03-20 09:05:00 +0000 UTC" firstStartedPulling="2026-03-20 09:05:01.735763814 +0000 UTC m=+468.421660188" lastFinishedPulling="2026-03-20 09:05:04.140928553 +0000 UTC m=+470.826824957" observedRunningTime="2026-03-20 09:05:04.796547062 +0000 UTC m=+471.482443436" watchObservedRunningTime="2026-03-20 09:05:04.802931256 +0000 UTC m=+471.488827630" Mar 20 09:05:04 crc kubenswrapper[4823]: I0320 09:05:04.818728 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2gqwq" podStartSLOduration=2.356307791 podStartE2EDuration="4.818711492s" podCreationTimestamp="2026-03-20 09:05:00 +0000 UTC" firstStartedPulling="2026-03-20 09:05:01.728354337 +0000 UTC m=+468.414250711" lastFinishedPulling="2026-03-20 09:05:04.190758028 +0000 UTC m=+470.876654412" observedRunningTime="2026-03-20 09:05:04.815570311 +0000 UTC m=+471.501466715" watchObservedRunningTime="2026-03-20 09:05:04.818711492 +0000 UTC m=+471.504607866" Mar 20 09:05:08 crc kubenswrapper[4823]: I0320 09:05:08.057363 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:05:08 crc kubenswrapper[4823]: I0320 09:05:08.057632 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:05:08 crc kubenswrapper[4823]: I0320 09:05:08.115351 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:05:08 crc kubenswrapper[4823]: I0320 09:05:08.246127 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:05:08 crc kubenswrapper[4823]: I0320 09:05:08.246295 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:05:08 crc kubenswrapper[4823]: I0320 09:05:08.290733 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:05:08 crc kubenswrapper[4823]: I0320 09:05:08.847406 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wzt57" Mar 20 09:05:08 crc kubenswrapper[4823]: I0320 09:05:08.857663 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j8d5w" Mar 20 09:05:09 crc kubenswrapper[4823]: I0320 09:05:09.429973 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:05:09 crc kubenswrapper[4823]: I0320 09:05:09.430070 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:05:09 crc kubenswrapper[4823]: I0320 09:05:09.430140 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:05:09 crc kubenswrapper[4823]: I0320 09:05:09.430847 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6d87d616f40330266a80a533f4cd60db866a189b08f07fd958207c63263d156f"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:05:09 crc kubenswrapper[4823]: I0320 09:05:09.430937 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://6d87d616f40330266a80a533f4cd60db866a189b08f07fd958207c63263d156f" gracePeriod=600 Mar 20 09:05:09 crc kubenswrapper[4823]: I0320 09:05:09.803194 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="6d87d616f40330266a80a533f4cd60db866a189b08f07fd958207c63263d156f" exitCode=0 Mar 20 09:05:09 crc kubenswrapper[4823]: I0320 09:05:09.803342 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"6d87d616f40330266a80a533f4cd60db866a189b08f07fd958207c63263d156f"} Mar 20 09:05:09 crc kubenswrapper[4823]: I0320 09:05:09.804289 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"684e65a34e103a8e275908b0f42a447fad127cb196d2c25b351518845765eb92"} Mar 20 09:05:09 crc kubenswrapper[4823]: I0320 09:05:09.804309 4823 scope.go:117] "RemoveContainer" containerID="9fbfc52ffb8bad407283d87fcef76ee5673ce80dbe11569cca221a7d11d223cb" Mar 20 09:05:10 crc kubenswrapper[4823]: I0320 09:05:10.452999 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:10 crc kubenswrapper[4823]: I0320 09:05:10.453358 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:10 crc kubenswrapper[4823]: I0320 09:05:10.520905 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:05:10 crc kubenswrapper[4823]: I0320 09:05:10.692010 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:10 crc kubenswrapper[4823]: I0320 09:05:10.692355 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:10 crc kubenswrapper[4823]: I0320 09:05:10.733683 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:10 crc kubenswrapper[4823]: I0320 09:05:10.866187 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2gqwq" Mar 20 09:05:10 crc kubenswrapper[4823]: I0320 09:05:10.866357 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.146006 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566626-z4qpc"] Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.147400 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566626-z4qpc" Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.150620 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.151398 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.151402 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.158486 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566626-z4qpc"] Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.248122 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpqtv\" (UniqueName: \"kubernetes.io/projected/0cfbe78d-6bcf-4fc2-b83e-e410c9af2522-kube-api-access-fpqtv\") pod \"auto-csr-approver-29566626-z4qpc\" (UID: \"0cfbe78d-6bcf-4fc2-b83e-e410c9af2522\") " pod="openshift-infra/auto-csr-approver-29566626-z4qpc" Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.350274 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpqtv\" (UniqueName: \"kubernetes.io/projected/0cfbe78d-6bcf-4fc2-b83e-e410c9af2522-kube-api-access-fpqtv\") pod \"auto-csr-approver-29566626-z4qpc\" (UID: \"0cfbe78d-6bcf-4fc2-b83e-e410c9af2522\") " pod="openshift-infra/auto-csr-approver-29566626-z4qpc" Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.369683 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpqtv\" (UniqueName: \"kubernetes.io/projected/0cfbe78d-6bcf-4fc2-b83e-e410c9af2522-kube-api-access-fpqtv\") pod \"auto-csr-approver-29566626-z4qpc\" (UID: \"0cfbe78d-6bcf-4fc2-b83e-e410c9af2522\") " pod="openshift-infra/auto-csr-approver-29566626-z4qpc" Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.475337 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566626-z4qpc" Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.663996 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566626-z4qpc"] Mar 20 09:06:00 crc kubenswrapper[4823]: I0320 09:06:00.671460 4823 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:06:01 crc kubenswrapper[4823]: I0320 09:06:01.120405 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566626-z4qpc" event={"ID":"0cfbe78d-6bcf-4fc2-b83e-e410c9af2522","Type":"ContainerStarted","Data":"e750c89e79cb3fa431ebcd7ca9c4cb9f42a033470044417deee85e6f4b3139fd"} Mar 20 09:06:02 crc kubenswrapper[4823]: I0320 09:06:02.129157 4823 generic.go:334] "Generic (PLEG): container finished" podID="0cfbe78d-6bcf-4fc2-b83e-e410c9af2522" containerID="02b48ffff6a5ae7706309c9d6a0252802210cdf945cf5637687d3ac055aa10df" exitCode=0 Mar 20 09:06:02 crc kubenswrapper[4823]: I0320 09:06:02.129219 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566626-z4qpc" event={"ID":"0cfbe78d-6bcf-4fc2-b83e-e410c9af2522","Type":"ContainerDied","Data":"02b48ffff6a5ae7706309c9d6a0252802210cdf945cf5637687d3ac055aa10df"} Mar 20 09:06:03 crc kubenswrapper[4823]: I0320 09:06:03.378222 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566626-z4qpc" Mar 20 09:06:03 crc kubenswrapper[4823]: I0320 09:06:03.388422 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpqtv\" (UniqueName: \"kubernetes.io/projected/0cfbe78d-6bcf-4fc2-b83e-e410c9af2522-kube-api-access-fpqtv\") pod \"0cfbe78d-6bcf-4fc2-b83e-e410c9af2522\" (UID: \"0cfbe78d-6bcf-4fc2-b83e-e410c9af2522\") " Mar 20 09:06:03 crc kubenswrapper[4823]: I0320 09:06:03.429852 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cfbe78d-6bcf-4fc2-b83e-e410c9af2522-kube-api-access-fpqtv" (OuterVolumeSpecName: "kube-api-access-fpqtv") pod "0cfbe78d-6bcf-4fc2-b83e-e410c9af2522" (UID: "0cfbe78d-6bcf-4fc2-b83e-e410c9af2522"). InnerVolumeSpecName "kube-api-access-fpqtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:06:03 crc kubenswrapper[4823]: I0320 09:06:03.489412 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpqtv\" (UniqueName: \"kubernetes.io/projected/0cfbe78d-6bcf-4fc2-b83e-e410c9af2522-kube-api-access-fpqtv\") on node \"crc\" DevicePath \"\"" Mar 20 09:06:04 crc kubenswrapper[4823]: I0320 09:06:04.147690 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566626-z4qpc" event={"ID":"0cfbe78d-6bcf-4fc2-b83e-e410c9af2522","Type":"ContainerDied","Data":"e750c89e79cb3fa431ebcd7ca9c4cb9f42a033470044417deee85e6f4b3139fd"} Mar 20 09:06:04 crc kubenswrapper[4823]: I0320 09:06:04.148065 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e750c89e79cb3fa431ebcd7ca9c4cb9f42a033470044417deee85e6f4b3139fd" Mar 20 09:06:04 crc kubenswrapper[4823]: I0320 09:06:04.147763 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566626-z4qpc" Mar 20 09:06:04 crc kubenswrapper[4823]: I0320 09:06:04.439555 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566620-4lfkt"] Mar 20 09:06:04 crc kubenswrapper[4823]: I0320 09:06:04.443567 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566620-4lfkt"] Mar 20 09:06:06 crc kubenswrapper[4823]: I0320 09:06:06.105803 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65a02b02-7f51-4774-b050-292c3fd4f70b" path="/var/lib/kubelet/pods/65a02b02-7f51-4774-b050-292c3fd4f70b/volumes" Mar 20 09:07:09 crc kubenswrapper[4823]: I0320 09:07:09.430278 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:07:09 crc kubenswrapper[4823]: I0320 09:07:09.431115 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:07:31 crc kubenswrapper[4823]: I0320 09:07:31.351072 4823 scope.go:117] "RemoveContainer" containerID="417b495fce674061bd7abe4a9f7535e74794235255d9303d98800d9ec849cc3c" Mar 20 09:07:39 crc kubenswrapper[4823]: I0320 09:07:39.430581 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:07:39 crc kubenswrapper[4823]: I0320 09:07:39.430931 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.143065 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566628-gq4lq"] Mar 20 09:08:00 crc kubenswrapper[4823]: E0320 09:08:00.144130 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfbe78d-6bcf-4fc2-b83e-e410c9af2522" containerName="oc" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.144146 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfbe78d-6bcf-4fc2-b83e-e410c9af2522" containerName="oc" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.144276 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfbe78d-6bcf-4fc2-b83e-e410c9af2522" containerName="oc" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.144673 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566628-gq4lq" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.147357 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.148403 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.148957 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.152121 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566628-gq4lq"] Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.235377 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxpkn\" (UniqueName: \"kubernetes.io/projected/7dc8b1fb-cbe3-4e67-8c08-878ac01d1853-kube-api-access-fxpkn\") pod \"auto-csr-approver-29566628-gq4lq\" (UID: \"7dc8b1fb-cbe3-4e67-8c08-878ac01d1853\") " pod="openshift-infra/auto-csr-approver-29566628-gq4lq" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.337427 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxpkn\" (UniqueName: \"kubernetes.io/projected/7dc8b1fb-cbe3-4e67-8c08-878ac01d1853-kube-api-access-fxpkn\") pod \"auto-csr-approver-29566628-gq4lq\" (UID: \"7dc8b1fb-cbe3-4e67-8c08-878ac01d1853\") " pod="openshift-infra/auto-csr-approver-29566628-gq4lq" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.359671 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxpkn\" (UniqueName: \"kubernetes.io/projected/7dc8b1fb-cbe3-4e67-8c08-878ac01d1853-kube-api-access-fxpkn\") pod \"auto-csr-approver-29566628-gq4lq\" (UID: \"7dc8b1fb-cbe3-4e67-8c08-878ac01d1853\") " pod="openshift-infra/auto-csr-approver-29566628-gq4lq" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.476778 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566628-gq4lq" Mar 20 09:08:00 crc kubenswrapper[4823]: I0320 09:08:00.665037 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566628-gq4lq"] Mar 20 09:08:01 crc kubenswrapper[4823]: I0320 09:08:00.924919 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566628-gq4lq" event={"ID":"7dc8b1fb-cbe3-4e67-8c08-878ac01d1853","Type":"ContainerStarted","Data":"8ce0f5a81e6d753bea73c39a5500f87c83862738c26285fb1c33a9adf033fc71"} Mar 20 09:08:02 crc kubenswrapper[4823]: I0320 09:08:02.942768 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566628-gq4lq" event={"ID":"7dc8b1fb-cbe3-4e67-8c08-878ac01d1853","Type":"ContainerStarted","Data":"1370ae68afa59b2b287edd65c55521ba880b60ace22809854da5ada82aaee09c"} Mar 20 09:08:02 crc kubenswrapper[4823]: I0320 09:08:02.966053 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566628-gq4lq" podStartSLOduration=1.132687949 podStartE2EDuration="2.966024238s" podCreationTimestamp="2026-03-20 09:08:00 +0000 UTC" firstStartedPulling="2026-03-20 09:08:00.675250897 +0000 UTC m=+647.361147291" lastFinishedPulling="2026-03-20 09:08:02.508587156 +0000 UTC m=+649.194483580" observedRunningTime="2026-03-20 09:08:02.961399906 +0000 UTC m=+649.647296340" watchObservedRunningTime="2026-03-20 09:08:02.966024238 +0000 UTC m=+649.651920652" Mar 20 09:08:03 crc kubenswrapper[4823]: I0320 09:08:03.953584 4823 generic.go:334] "Generic (PLEG): container finished" podID="7dc8b1fb-cbe3-4e67-8c08-878ac01d1853" containerID="1370ae68afa59b2b287edd65c55521ba880b60ace22809854da5ada82aaee09c" exitCode=0 Mar 20 09:08:03 crc kubenswrapper[4823]: I0320 09:08:03.953642 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566628-gq4lq" event={"ID":"7dc8b1fb-cbe3-4e67-8c08-878ac01d1853","Type":"ContainerDied","Data":"1370ae68afa59b2b287edd65c55521ba880b60ace22809854da5ada82aaee09c"} Mar 20 09:08:05 crc kubenswrapper[4823]: I0320 09:08:05.206135 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566628-gq4lq" Mar 20 09:08:05 crc kubenswrapper[4823]: I0320 09:08:05.302215 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxpkn\" (UniqueName: \"kubernetes.io/projected/7dc8b1fb-cbe3-4e67-8c08-878ac01d1853-kube-api-access-fxpkn\") pod \"7dc8b1fb-cbe3-4e67-8c08-878ac01d1853\" (UID: \"7dc8b1fb-cbe3-4e67-8c08-878ac01d1853\") " Mar 20 09:08:05 crc kubenswrapper[4823]: I0320 09:08:05.311541 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dc8b1fb-cbe3-4e67-8c08-878ac01d1853-kube-api-access-fxpkn" (OuterVolumeSpecName: "kube-api-access-fxpkn") pod "7dc8b1fb-cbe3-4e67-8c08-878ac01d1853" (UID: "7dc8b1fb-cbe3-4e67-8c08-878ac01d1853"). InnerVolumeSpecName "kube-api-access-fxpkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:08:05 crc kubenswrapper[4823]: I0320 09:08:05.404040 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxpkn\" (UniqueName: \"kubernetes.io/projected/7dc8b1fb-cbe3-4e67-8c08-878ac01d1853-kube-api-access-fxpkn\") on node \"crc\" DevicePath \"\"" Mar 20 09:08:05 crc kubenswrapper[4823]: I0320 09:08:05.968396 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566628-gq4lq" event={"ID":"7dc8b1fb-cbe3-4e67-8c08-878ac01d1853","Type":"ContainerDied","Data":"8ce0f5a81e6d753bea73c39a5500f87c83862738c26285fb1c33a9adf033fc71"} Mar 20 09:08:05 crc kubenswrapper[4823]: I0320 09:08:05.968438 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ce0f5a81e6d753bea73c39a5500f87c83862738c26285fb1c33a9adf033fc71" Mar 20 09:08:05 crc kubenswrapper[4823]: I0320 09:08:05.968504 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566628-gq4lq" Mar 20 09:08:06 crc kubenswrapper[4823]: I0320 09:08:06.283998 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566622-nc9qz"] Mar 20 09:08:06 crc kubenswrapper[4823]: I0320 09:08:06.291986 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566622-nc9qz"] Mar 20 09:08:08 crc kubenswrapper[4823]: I0320 09:08:08.110042 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56782166-f1af-439a-b19b-8eb59caaffe9" path="/var/lib/kubelet/pods/56782166-f1af-439a-b19b-8eb59caaffe9/volumes" Mar 20 09:08:09 crc kubenswrapper[4823]: I0320 09:08:09.430317 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:08:09 crc kubenswrapper[4823]: I0320 09:08:09.430372 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:08:09 crc kubenswrapper[4823]: I0320 09:08:09.430415 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:08:09 crc kubenswrapper[4823]: I0320 09:08:09.430894 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"684e65a34e103a8e275908b0f42a447fad127cb196d2c25b351518845765eb92"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:08:09 crc kubenswrapper[4823]: I0320 09:08:09.430962 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://684e65a34e103a8e275908b0f42a447fad127cb196d2c25b351518845765eb92" gracePeriod=600 Mar 20 09:08:09 crc kubenswrapper[4823]: I0320 09:08:09.997493 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="684e65a34e103a8e275908b0f42a447fad127cb196d2c25b351518845765eb92" exitCode=0 Mar 20 09:08:09 crc kubenswrapper[4823]: I0320 09:08:09.997580 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"684e65a34e103a8e275908b0f42a447fad127cb196d2c25b351518845765eb92"} Mar 20 09:08:09 crc kubenswrapper[4823]: I0320 09:08:09.997916 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"60ec5c90c519363b5a41f87bdb8567fcf49b737ce1cfeea150ceae7849bf0158"} Mar 20 09:08:09 crc kubenswrapper[4823]: I0320 09:08:09.997949 4823 scope.go:117] "RemoveContainer" containerID="6d87d616f40330266a80a533f4cd60db866a189b08f07fd958207c63263d156f" Mar 20 09:08:31 crc kubenswrapper[4823]: I0320 09:08:31.409596 4823 scope.go:117] "RemoveContainer" containerID="6f4275c05d7506341508bc9cef63cbad8af2db169cebd75b765c564d99f7f698" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.341183 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6m2ss"] Mar 20 09:08:36 crc kubenswrapper[4823]: E0320 09:08:36.341700 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc8b1fb-cbe3-4e67-8c08-878ac01d1853" containerName="oc" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.341715 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc8b1fb-cbe3-4e67-8c08-878ac01d1853" containerName="oc" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.341842 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dc8b1fb-cbe3-4e67-8c08-878ac01d1853" containerName="oc" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.342286 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.354511 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6m2ss"] Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.442076 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b2ca188f-61d9-4390-9f92-6175a49fef20-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.442144 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b2ca188f-61d9-4390-9f92-6175a49fef20-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.442172 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b2ca188f-61d9-4390-9f92-6175a49fef20-registry-tls\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.442217 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.442342 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b2ca188f-61d9-4390-9f92-6175a49fef20-registry-certificates\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.442436 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b2ca188f-61d9-4390-9f92-6175a49fef20-bound-sa-token\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.442475 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b2ca188f-61d9-4390-9f92-6175a49fef20-trusted-ca\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.442506 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w4sq\" (UniqueName: \"kubernetes.io/projected/b2ca188f-61d9-4390-9f92-6175a49fef20-kube-api-access-6w4sq\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.467418 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.543648 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b2ca188f-61d9-4390-9f92-6175a49fef20-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.543708 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b2ca188f-61d9-4390-9f92-6175a49fef20-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.543755 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b2ca188f-61d9-4390-9f92-6175a49fef20-registry-tls\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.543805 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b2ca188f-61d9-4390-9f92-6175a49fef20-registry-certificates\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.543836 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b2ca188f-61d9-4390-9f92-6175a49fef20-bound-sa-token\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.543858 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b2ca188f-61d9-4390-9f92-6175a49fef20-trusted-ca\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.543878 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w4sq\" (UniqueName: \"kubernetes.io/projected/b2ca188f-61d9-4390-9f92-6175a49fef20-kube-api-access-6w4sq\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.544373 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b2ca188f-61d9-4390-9f92-6175a49fef20-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.545361 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b2ca188f-61d9-4390-9f92-6175a49fef20-trusted-ca\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.545711 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b2ca188f-61d9-4390-9f92-6175a49fef20-registry-certificates\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.549035 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b2ca188f-61d9-4390-9f92-6175a49fef20-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.549877 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b2ca188f-61d9-4390-9f92-6175a49fef20-registry-tls\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.567393 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b2ca188f-61d9-4390-9f92-6175a49fef20-bound-sa-token\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.581277 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w4sq\" (UniqueName: \"kubernetes.io/projected/b2ca188f-61d9-4390-9f92-6175a49fef20-kube-api-access-6w4sq\") pod \"image-registry-66df7c8f76-6m2ss\" (UID: \"b2ca188f-61d9-4390-9f92-6175a49fef20\") " pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.685353 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:36 crc kubenswrapper[4823]: I0320 09:08:36.917050 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6m2ss"] Mar 20 09:08:37 crc kubenswrapper[4823]: I0320 09:08:37.175839 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" event={"ID":"b2ca188f-61d9-4390-9f92-6175a49fef20","Type":"ContainerStarted","Data":"c663945db708e8765cf0abb16a08ca05c3c5ed21edebe3d1c26befb22a0b1656"} Mar 20 09:08:37 crc kubenswrapper[4823]: I0320 09:08:37.176026 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" event={"ID":"b2ca188f-61d9-4390-9f92-6175a49fef20","Type":"ContainerStarted","Data":"4a71b4d3403482ef042ccb82fff02ff3a9faaa94458f50c32e2e1820105b85b4"} Mar 20 09:08:37 crc kubenswrapper[4823]: I0320 09:08:37.176640 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:37 crc kubenswrapper[4823]: I0320 09:08:37.209422 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" podStartSLOduration=1.2094039269999999 podStartE2EDuration="1.209403927s" podCreationTimestamp="2026-03-20 09:08:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:08:37.206513818 +0000 UTC m=+683.892410192" watchObservedRunningTime="2026-03-20 09:08:37.209403927 +0000 UTC m=+683.895300301" Mar 20 09:08:56 crc kubenswrapper[4823]: I0320 09:08:56.694682 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-6m2ss" Mar 20 09:08:56 crc kubenswrapper[4823]: I0320 09:08:56.770983 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwrsf"] Mar 20 09:09:21 crc kubenswrapper[4823]: I0320 09:09:21.820047 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" podUID="e9000667-ed2e-4fea-b03f-c9760e59a604" containerName="registry" containerID="cri-o://d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f" gracePeriod=30 Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.187832 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.307862 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-trusted-ca\") pod \"e9000667-ed2e-4fea-b03f-c9760e59a604\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.307945 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-tls\") pod \"e9000667-ed2e-4fea-b03f-c9760e59a604\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.308004 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e9000667-ed2e-4fea-b03f-c9760e59a604-ca-trust-extracted\") pod \"e9000667-ed2e-4fea-b03f-c9760e59a604\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.308059 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-certificates\") pod \"e9000667-ed2e-4fea-b03f-c9760e59a604\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.308137 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6nnp\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-kube-api-access-n6nnp\") pod \"e9000667-ed2e-4fea-b03f-c9760e59a604\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.308208 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-bound-sa-token\") pod \"e9000667-ed2e-4fea-b03f-c9760e59a604\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.308498 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"e9000667-ed2e-4fea-b03f-c9760e59a604\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.309666 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "e9000667-ed2e-4fea-b03f-c9760e59a604" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.310100 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e9000667-ed2e-4fea-b03f-c9760e59a604-installation-pull-secrets\") pod \"e9000667-ed2e-4fea-b03f-c9760e59a604\" (UID: \"e9000667-ed2e-4fea-b03f-c9760e59a604\") " Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.310509 4823 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.311646 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "e9000667-ed2e-4fea-b03f-c9760e59a604" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.316002 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9000667-ed2e-4fea-b03f-c9760e59a604-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "e9000667-ed2e-4fea-b03f-c9760e59a604" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.317040 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "e9000667-ed2e-4fea-b03f-c9760e59a604" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.318516 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "e9000667-ed2e-4fea-b03f-c9760e59a604" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.319881 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-kube-api-access-n6nnp" (OuterVolumeSpecName: "kube-api-access-n6nnp") pod "e9000667-ed2e-4fea-b03f-c9760e59a604" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604"). InnerVolumeSpecName "kube-api-access-n6nnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.322366 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "e9000667-ed2e-4fea-b03f-c9760e59a604" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.337866 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9000667-ed2e-4fea-b03f-c9760e59a604-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "e9000667-ed2e-4fea-b03f-c9760e59a604" (UID: "e9000667-ed2e-4fea-b03f-c9760e59a604"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.411680 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6nnp\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-kube-api-access-n6nnp\") on node \"crc\" DevicePath \"\"" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.411785 4823 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.411815 4823 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e9000667-ed2e-4fea-b03f-c9760e59a604-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.411840 4823 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9000667-ed2e-4fea-b03f-c9760e59a604-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.411869 4823 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e9000667-ed2e-4fea-b03f-c9760e59a604-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.411893 4823 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e9000667-ed2e-4fea-b03f-c9760e59a604-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.531265 4823 generic.go:334] "Generic (PLEG): container finished" podID="e9000667-ed2e-4fea-b03f-c9760e59a604" containerID="d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f" exitCode=0 Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.531358 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" event={"ID":"e9000667-ed2e-4fea-b03f-c9760e59a604","Type":"ContainerDied","Data":"d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f"} Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.531409 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" event={"ID":"e9000667-ed2e-4fea-b03f-c9760e59a604","Type":"ContainerDied","Data":"fb4a34145e231b8f132c9acfcac8fd1d470eec42959da4548ad7ea324f91e57c"} Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.531424 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fwrsf" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.531448 4823 scope.go:117] "RemoveContainer" containerID="d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.561630 4823 scope.go:117] "RemoveContainer" containerID="d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f" Mar 20 09:09:22 crc kubenswrapper[4823]: E0320 09:09:22.562552 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f\": container with ID starting with d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f not found: ID does not exist" containerID="d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.562610 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f"} err="failed to get container status \"d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f\": rpc error: code = NotFound desc = could not find container \"d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f\": container with ID starting with d881a9cc61f7f354abb3d673f799a4c791d3726c5f155127270aef1716ff311f not found: ID does not exist" Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.581127 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwrsf"] Mar 20 09:09:22 crc kubenswrapper[4823]: I0320 09:09:22.584571 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwrsf"] Mar 20 09:09:24 crc kubenswrapper[4823]: I0320 09:09:24.108372 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9000667-ed2e-4fea-b03f-c9760e59a604" path="/var/lib/kubelet/pods/e9000667-ed2e-4fea-b03f-c9760e59a604/volumes" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.144820 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566630-r54tb"] Mar 20 09:10:00 crc kubenswrapper[4823]: E0320 09:10:00.146606 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9000667-ed2e-4fea-b03f-c9760e59a604" containerName="registry" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.146626 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9000667-ed2e-4fea-b03f-c9760e59a604" containerName="registry" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.146809 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9000667-ed2e-4fea-b03f-c9760e59a604" containerName="registry" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.147330 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566630-r54tb" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.149681 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566630-r54tb"] Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.151357 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.151435 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.151724 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.271439 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h546\" (UniqueName: \"kubernetes.io/projected/e3431f35-1655-4dae-89d3-d30c2a5065bb-kube-api-access-2h546\") pod \"auto-csr-approver-29566630-r54tb\" (UID: \"e3431f35-1655-4dae-89d3-d30c2a5065bb\") " pod="openshift-infra/auto-csr-approver-29566630-r54tb" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.373551 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h546\" (UniqueName: \"kubernetes.io/projected/e3431f35-1655-4dae-89d3-d30c2a5065bb-kube-api-access-2h546\") pod \"auto-csr-approver-29566630-r54tb\" (UID: \"e3431f35-1655-4dae-89d3-d30c2a5065bb\") " pod="openshift-infra/auto-csr-approver-29566630-r54tb" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.407854 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h546\" (UniqueName: \"kubernetes.io/projected/e3431f35-1655-4dae-89d3-d30c2a5065bb-kube-api-access-2h546\") pod \"auto-csr-approver-29566630-r54tb\" (UID: \"e3431f35-1655-4dae-89d3-d30c2a5065bb\") " pod="openshift-infra/auto-csr-approver-29566630-r54tb" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.476466 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566630-r54tb" Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.718825 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566630-r54tb"] Mar 20 09:10:00 crc kubenswrapper[4823]: I0320 09:10:00.770693 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566630-r54tb" event={"ID":"e3431f35-1655-4dae-89d3-d30c2a5065bb","Type":"ContainerStarted","Data":"87dd4c3e9740c6dec67bb3a406625fc43726cc5860d5ec9ca752216fa30f7f2a"} Mar 20 09:10:02 crc kubenswrapper[4823]: I0320 09:10:02.783101 4823 generic.go:334] "Generic (PLEG): container finished" podID="e3431f35-1655-4dae-89d3-d30c2a5065bb" containerID="e4023c35ef355eac2c6cc566acf642f1f05b82742e42861f1a78fd51ff85e10d" exitCode=0 Mar 20 09:10:02 crc kubenswrapper[4823]: I0320 09:10:02.783186 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566630-r54tb" event={"ID":"e3431f35-1655-4dae-89d3-d30c2a5065bb","Type":"ContainerDied","Data":"e4023c35ef355eac2c6cc566acf642f1f05b82742e42861f1a78fd51ff85e10d"} Mar 20 09:10:03 crc kubenswrapper[4823]: I0320 09:10:03.973291 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566630-r54tb" Mar 20 09:10:04 crc kubenswrapper[4823]: I0320 09:10:04.119755 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h546\" (UniqueName: \"kubernetes.io/projected/e3431f35-1655-4dae-89d3-d30c2a5065bb-kube-api-access-2h546\") pod \"e3431f35-1655-4dae-89d3-d30c2a5065bb\" (UID: \"e3431f35-1655-4dae-89d3-d30c2a5065bb\") " Mar 20 09:10:04 crc kubenswrapper[4823]: I0320 09:10:04.126157 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3431f35-1655-4dae-89d3-d30c2a5065bb-kube-api-access-2h546" (OuterVolumeSpecName: "kube-api-access-2h546") pod "e3431f35-1655-4dae-89d3-d30c2a5065bb" (UID: "e3431f35-1655-4dae-89d3-d30c2a5065bb"). InnerVolumeSpecName "kube-api-access-2h546". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:10:04 crc kubenswrapper[4823]: I0320 09:10:04.221664 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h546\" (UniqueName: \"kubernetes.io/projected/e3431f35-1655-4dae-89d3-d30c2a5065bb-kube-api-access-2h546\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:04 crc kubenswrapper[4823]: I0320 09:10:04.796525 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566630-r54tb" event={"ID":"e3431f35-1655-4dae-89d3-d30c2a5065bb","Type":"ContainerDied","Data":"87dd4c3e9740c6dec67bb3a406625fc43726cc5860d5ec9ca752216fa30f7f2a"} Mar 20 09:10:04 crc kubenswrapper[4823]: I0320 09:10:04.796584 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87dd4c3e9740c6dec67bb3a406625fc43726cc5860d5ec9ca752216fa30f7f2a" Mar 20 09:10:04 crc kubenswrapper[4823]: I0320 09:10:04.796603 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566630-r54tb" Mar 20 09:10:05 crc kubenswrapper[4823]: I0320 09:10:05.025369 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566624-7h65b"] Mar 20 09:10:05 crc kubenswrapper[4823]: I0320 09:10:05.029706 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566624-7h65b"] Mar 20 09:10:06 crc kubenswrapper[4823]: I0320 09:10:06.109035 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2da31dce-84ef-4cc4-a959-97cbc515f685" path="/var/lib/kubelet/pods/2da31dce-84ef-4cc4-a959-97cbc515f685/volumes" Mar 20 09:10:09 crc kubenswrapper[4823]: I0320 09:10:09.430519 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:10:09 crc kubenswrapper[4823]: I0320 09:10:09.430581 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.945645 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-hfnfq"] Mar 20 09:10:10 crc kubenswrapper[4823]: E0320 09:10:10.946155 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3431f35-1655-4dae-89d3-d30c2a5065bb" containerName="oc" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.946178 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3431f35-1655-4dae-89d3-d30c2a5065bb" containerName="oc" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.946275 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3431f35-1655-4dae-89d3-d30c2a5065bb" containerName="oc" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.946610 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hfnfq" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.948506 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.948539 4823 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-cp7l7" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.948871 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.951621 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-nkqmf"] Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.952136 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-nkqmf" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.953617 4823 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-nr5bh" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.962169 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-hfnfq"] Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.969034 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-8jmq6"] Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.969915 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-8jmq6" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.975131 4823 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-nhzpb" Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.982864 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-nkqmf"] Mar 20 09:10:10 crc kubenswrapper[4823]: I0320 09:10:10.990051 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-8jmq6"] Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.104767 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncmgt\" (UniqueName: \"kubernetes.io/projected/d65f0664-ecd0-4123-81e3-39ec448311d4-kube-api-access-ncmgt\") pod \"cert-manager-webhook-687f57d79b-8jmq6\" (UID: \"d65f0664-ecd0-4123-81e3-39ec448311d4\") " pod="cert-manager/cert-manager-webhook-687f57d79b-8jmq6" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.104832 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82v87\" (UniqueName: \"kubernetes.io/projected/393e0c29-456f-4547-a8fd-7f14e3b5285c-kube-api-access-82v87\") pod \"cert-manager-cainjector-cf98fcc89-hfnfq\" (UID: \"393e0c29-456f-4547-a8fd-7f14e3b5285c\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-hfnfq" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.104891 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6pjz\" (UniqueName: \"kubernetes.io/projected/db934289-d3d9-4319-a56f-37faa9181e58-kube-api-access-g6pjz\") pod \"cert-manager-858654f9db-nkqmf\" (UID: \"db934289-d3d9-4319-a56f-37faa9181e58\") " pod="cert-manager/cert-manager-858654f9db-nkqmf" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.206446 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6pjz\" (UniqueName: \"kubernetes.io/projected/db934289-d3d9-4319-a56f-37faa9181e58-kube-api-access-g6pjz\") pod \"cert-manager-858654f9db-nkqmf\" (UID: \"db934289-d3d9-4319-a56f-37faa9181e58\") " pod="cert-manager/cert-manager-858654f9db-nkqmf" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.206503 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncmgt\" (UniqueName: \"kubernetes.io/projected/d65f0664-ecd0-4123-81e3-39ec448311d4-kube-api-access-ncmgt\") pod \"cert-manager-webhook-687f57d79b-8jmq6\" (UID: \"d65f0664-ecd0-4123-81e3-39ec448311d4\") " pod="cert-manager/cert-manager-webhook-687f57d79b-8jmq6" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.206549 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82v87\" (UniqueName: \"kubernetes.io/projected/393e0c29-456f-4547-a8fd-7f14e3b5285c-kube-api-access-82v87\") pod \"cert-manager-cainjector-cf98fcc89-hfnfq\" (UID: \"393e0c29-456f-4547-a8fd-7f14e3b5285c\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-hfnfq" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.224551 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82v87\" (UniqueName: \"kubernetes.io/projected/393e0c29-456f-4547-a8fd-7f14e3b5285c-kube-api-access-82v87\") pod \"cert-manager-cainjector-cf98fcc89-hfnfq\" (UID: \"393e0c29-456f-4547-a8fd-7f14e3b5285c\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-hfnfq" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.225449 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6pjz\" (UniqueName: \"kubernetes.io/projected/db934289-d3d9-4319-a56f-37faa9181e58-kube-api-access-g6pjz\") pod \"cert-manager-858654f9db-nkqmf\" (UID: \"db934289-d3d9-4319-a56f-37faa9181e58\") " pod="cert-manager/cert-manager-858654f9db-nkqmf" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.226459 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncmgt\" (UniqueName: \"kubernetes.io/projected/d65f0664-ecd0-4123-81e3-39ec448311d4-kube-api-access-ncmgt\") pod \"cert-manager-webhook-687f57d79b-8jmq6\" (UID: \"d65f0664-ecd0-4123-81e3-39ec448311d4\") " pod="cert-manager/cert-manager-webhook-687f57d79b-8jmq6" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.267410 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hfnfq" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.275829 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-nkqmf" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.288234 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-8jmq6" Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.500551 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-hfnfq"] Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.747188 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-8jmq6"] Mar 20 09:10:11 crc kubenswrapper[4823]: W0320 09:10:11.751100 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd65f0664_ecd0_4123_81e3_39ec448311d4.slice/crio-7ac4a84d92b159f49cea294d3f73681bdbb8d18a4453b397f26ff88d79fba39b WatchSource:0}: Error finding container 7ac4a84d92b159f49cea294d3f73681bdbb8d18a4453b397f26ff88d79fba39b: Status 404 returned error can't find the container with id 7ac4a84d92b159f49cea294d3f73681bdbb8d18a4453b397f26ff88d79fba39b Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.755154 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-nkqmf"] Mar 20 09:10:11 crc kubenswrapper[4823]: W0320 09:10:11.760576 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb934289_d3d9_4319_a56f_37faa9181e58.slice/crio-3fbdcd9b47d7ddc759ee671ca2222cbb0cc9bca58c3eb1c76cd32d4728e96e37 WatchSource:0}: Error finding container 3fbdcd9b47d7ddc759ee671ca2222cbb0cc9bca58c3eb1c76cd32d4728e96e37: Status 404 returned error can't find the container with id 3fbdcd9b47d7ddc759ee671ca2222cbb0cc9bca58c3eb1c76cd32d4728e96e37 Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.833549 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hfnfq" event={"ID":"393e0c29-456f-4547-a8fd-7f14e3b5285c","Type":"ContainerStarted","Data":"30d2547e2aafaa4c3cf456776b71c68e96617aa68d9eb57ec4edeca0f17b3eb4"} Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.834488 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-8jmq6" event={"ID":"d65f0664-ecd0-4123-81e3-39ec448311d4","Type":"ContainerStarted","Data":"7ac4a84d92b159f49cea294d3f73681bdbb8d18a4453b397f26ff88d79fba39b"} Mar 20 09:10:11 crc kubenswrapper[4823]: I0320 09:10:11.835233 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-nkqmf" event={"ID":"db934289-d3d9-4319-a56f-37faa9181e58","Type":"ContainerStarted","Data":"3fbdcd9b47d7ddc759ee671ca2222cbb0cc9bca58c3eb1c76cd32d4728e96e37"} Mar 20 09:10:14 crc kubenswrapper[4823]: I0320 09:10:14.853379 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-8jmq6" event={"ID":"d65f0664-ecd0-4123-81e3-39ec448311d4","Type":"ContainerStarted","Data":"759d79fc3a5e7816c8f13ef5db1663e80c5c203a393e07e87dc965288b77d342"} Mar 20 09:10:14 crc kubenswrapper[4823]: I0320 09:10:14.854004 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-8jmq6" Mar 20 09:10:14 crc kubenswrapper[4823]: I0320 09:10:14.855695 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hfnfq" event={"ID":"393e0c29-456f-4547-a8fd-7f14e3b5285c","Type":"ContainerStarted","Data":"01be4162d17b574fdbd0112fbe9de5019fe7bdf7827cd7d55a1cb6f7054a102d"} Mar 20 09:10:14 crc kubenswrapper[4823]: I0320 09:10:14.874369 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-8jmq6" podStartSLOduration=2.153725067 podStartE2EDuration="4.87434695s" podCreationTimestamp="2026-03-20 09:10:10 +0000 UTC" firstStartedPulling="2026-03-20 09:10:11.754065463 +0000 UTC m=+778.439961837" lastFinishedPulling="2026-03-20 09:10:14.474687346 +0000 UTC m=+781.160583720" observedRunningTime="2026-03-20 09:10:14.866719862 +0000 UTC m=+781.552616236" watchObservedRunningTime="2026-03-20 09:10:14.87434695 +0000 UTC m=+781.560243324" Mar 20 09:10:14 crc kubenswrapper[4823]: I0320 09:10:14.883704 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hfnfq" podStartSLOduration=1.923097486 podStartE2EDuration="4.883686559s" podCreationTimestamp="2026-03-20 09:10:10 +0000 UTC" firstStartedPulling="2026-03-20 09:10:11.513898518 +0000 UTC m=+778.199794882" lastFinishedPulling="2026-03-20 09:10:14.474487581 +0000 UTC m=+781.160383955" observedRunningTime="2026-03-20 09:10:14.883244886 +0000 UTC m=+781.569141310" watchObservedRunningTime="2026-03-20 09:10:14.883686559 +0000 UTC m=+781.569582933" Mar 20 09:10:15 crc kubenswrapper[4823]: I0320 09:10:15.862025 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-nkqmf" event={"ID":"db934289-d3d9-4319-a56f-37faa9181e58","Type":"ContainerStarted","Data":"f299b2819ef643ae924be55d8926d349587df90407be18336de4bc1ade056aa7"} Mar 20 09:10:15 crc kubenswrapper[4823]: I0320 09:10:15.880625 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-nkqmf" podStartSLOduration=2.060330301 podStartE2EDuration="5.880601376s" podCreationTimestamp="2026-03-20 09:10:10 +0000 UTC" firstStartedPulling="2026-03-20 09:10:11.762549486 +0000 UTC m=+778.448445860" lastFinishedPulling="2026-03-20 09:10:15.582820561 +0000 UTC m=+782.268716935" observedRunningTime="2026-03-20 09:10:15.875840728 +0000 UTC m=+782.561737122" watchObservedRunningTime="2026-03-20 09:10:15.880601376 +0000 UTC m=+782.566497750" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.080063 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2x54k"] Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.081183 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="northd" containerID="cri-o://bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f" gracePeriod=30 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.081259 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="kube-rbac-proxy-node" containerID="cri-o://46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615" gracePeriod=30 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.081406 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="sbdb" containerID="cri-o://4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12" gracePeriod=30 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.081458 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c" gracePeriod=30 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.081465 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="nbdb" containerID="cri-o://2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55" gracePeriod=30 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.081543 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovn-acl-logging" containerID="cri-o://37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680" gracePeriod=30 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.081785 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovn-controller" containerID="cri-o://8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3" gracePeriod=30 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.118251 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" containerID="cri-o://9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1" gracePeriod=30 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.290815 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-8jmq6" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.423886 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/3.log" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.427924 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovn-acl-logging/0.log" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.429546 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovn-controller/0.log" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.430192 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.496717 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6qll5"] Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.496994 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497008 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497019 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="northd" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497025 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="northd" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497033 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="nbdb" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497039 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="nbdb" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497046 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497052 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497058 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="kube-rbac-proxy-node" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497064 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="kube-rbac-proxy-node" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497073 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovn-acl-logging" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497080 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovn-acl-logging" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497089 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497094 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497102 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497107 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497115 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497120 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497130 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="kubecfg-setup" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497136 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="kubecfg-setup" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497143 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovn-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497149 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovn-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497158 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="sbdb" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497163 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="sbdb" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497263 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="nbdb" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497275 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497287 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="sbdb" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497298 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovn-acl-logging" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497307 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497314 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="kube-rbac-proxy-node" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497323 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="northd" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497335 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497344 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497351 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovn-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.497433 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497440 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497527 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.497537 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerName="ovnkube-controller" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.499147 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548051 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548107 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-log-socket\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548151 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-slash\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548187 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-env-overrides\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548208 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-kubelet\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548232 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-ovn\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548271 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-bin\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548299 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovn-node-metrics-cert\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548316 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-node-log\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548335 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-config\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548353 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-etc-openvswitch\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548396 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-systemd-units\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548418 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-script-lib\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548452 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-ovn-kubernetes\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548472 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-var-lib-openvswitch\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548496 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-systemd\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548523 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-netd\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548553 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-openvswitch\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548576 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-netns\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.548596 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwkjs\" (UniqueName: \"kubernetes.io/projected/1bed1862-5bcd-46fe-bc84-2dbab416086b-kube-api-access-mwkjs\") pod \"1bed1862-5bcd-46fe-bc84-2dbab416086b\" (UID: \"1bed1862-5bcd-46fe-bc84-2dbab416086b\") " Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.550954 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551029 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551053 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-log-socket" (OuterVolumeSpecName: "log-socket") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551077 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-slash" (OuterVolumeSpecName: "host-slash") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551142 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551248 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551280 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551292 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551328 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551346 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551466 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551513 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551555 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-node-log" (OuterVolumeSpecName: "node-log") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551657 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551810 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551860 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.551883 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.556870 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.557237 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bed1862-5bcd-46fe-bc84-2dbab416086b-kube-api-access-mwkjs" (OuterVolumeSpecName: "kube-api-access-mwkjs") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "kube-api-access-mwkjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.566656 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "1bed1862-5bcd-46fe-bc84-2dbab416086b" (UID: "1bed1862-5bcd-46fe-bc84-2dbab416086b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.649992 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/533a29bc-33f6-4b3c-b222-d601beb87ab8-ovn-node-metrics-cert\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650038 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/533a29bc-33f6-4b3c-b222-d601beb87ab8-ovnkube-config\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650184 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-run-netns\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650287 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/533a29bc-33f6-4b3c-b222-d601beb87ab8-env-overrides\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650346 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-run-ovn-kubernetes\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650384 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-cni-bin\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650444 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-systemd-units\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650487 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-node-log\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650551 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98ndw\" (UniqueName: \"kubernetes.io/projected/533a29bc-33f6-4b3c-b222-d601beb87ab8-kube-api-access-98ndw\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650592 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-run-ovn\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650629 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-kubelet\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650669 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-etc-openvswitch\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650842 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.650993 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-cni-netd\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651111 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-run-openvswitch\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651218 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-run-systemd\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651311 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/533a29bc-33f6-4b3c-b222-d601beb87ab8-ovnkube-script-lib\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651520 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-var-lib-openvswitch\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651631 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-log-socket\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651675 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-slash\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651859 4823 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651888 4823 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651913 4823 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651934 4823 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651953 4823 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651972 4823 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-node-log\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.651990 4823 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652009 4823 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652027 4823 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1bed1862-5bcd-46fe-bc84-2dbab416086b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652045 4823 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652064 4823 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652083 4823 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652101 4823 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652118 4823 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652137 4823 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652154 4823 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652172 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwkjs\" (UniqueName: \"kubernetes.io/projected/1bed1862-5bcd-46fe-bc84-2dbab416086b-kube-api-access-mwkjs\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652193 4823 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-log-socket\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652212 4823 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.652236 4823 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1bed1862-5bcd-46fe-bc84-2dbab416086b-host-slash\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753105 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98ndw\" (UniqueName: \"kubernetes.io/projected/533a29bc-33f6-4b3c-b222-d601beb87ab8-kube-api-access-98ndw\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753175 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-run-ovn\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753343 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-run-ovn\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753402 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-kubelet\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753492 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-kubelet\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753531 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-etc-openvswitch\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753569 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753620 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-etc-openvswitch\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753693 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753729 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-cni-netd\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753782 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-cni-netd\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753829 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-run-openvswitch\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753848 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-run-openvswitch\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753869 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-run-systemd\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753902 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/533a29bc-33f6-4b3c-b222-d601beb87ab8-ovnkube-script-lib\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753952 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-var-lib-openvswitch\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.753985 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-log-socket\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754014 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-slash\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754023 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-run-systemd\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754113 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-slash\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754129 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-log-socket\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754076 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-var-lib-openvswitch\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754080 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/533a29bc-33f6-4b3c-b222-d601beb87ab8-ovn-node-metrics-cert\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754216 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/533a29bc-33f6-4b3c-b222-d601beb87ab8-ovnkube-config\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754254 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-run-netns\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754296 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/533a29bc-33f6-4b3c-b222-d601beb87ab8-env-overrides\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754333 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-run-ovn-kubernetes\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754364 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-cni-bin\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754398 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-systemd-units\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754425 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-node-log\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754561 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-node-log\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754626 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-run-netns\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754673 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/533a29bc-33f6-4b3c-b222-d601beb87ab8-ovnkube-script-lib\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754717 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-cni-bin\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754765 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-systemd-units\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.754781 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/533a29bc-33f6-4b3c-b222-d601beb87ab8-host-run-ovn-kubernetes\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.755440 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/533a29bc-33f6-4b3c-b222-d601beb87ab8-env-overrides\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.755801 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/533a29bc-33f6-4b3c-b222-d601beb87ab8-ovnkube-config\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.757987 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/533a29bc-33f6-4b3c-b222-d601beb87ab8-ovn-node-metrics-cert\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.785544 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98ndw\" (UniqueName: \"kubernetes.io/projected/533a29bc-33f6-4b3c-b222-d601beb87ab8-kube-api-access-98ndw\") pod \"ovnkube-node-6qll5\" (UID: \"533a29bc-33f6-4b3c-b222-d601beb87ab8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.811131 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:21 crc kubenswrapper[4823]: W0320 09:10:21.846997 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod533a29bc_33f6_4b3c_b222_d601beb87ab8.slice/crio-ae266d7019a1bd19fb9b757de9ed1b30e474a95df0c927e1b7745d6904357986 WatchSource:0}: Error finding container ae266d7019a1bd19fb9b757de9ed1b30e474a95df0c927e1b7745d6904357986: Status 404 returned error can't find the container with id ae266d7019a1bd19fb9b757de9ed1b30e474a95df0c927e1b7745d6904357986 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.906091 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9vsj7_d33f9d63-34d7-4722-95d5-13bdcb32b99a/kube-multus/2.log" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.906773 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9vsj7_d33f9d63-34d7-4722-95d5-13bdcb32b99a/kube-multus/1.log" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.906853 4823 generic.go:334] "Generic (PLEG): container finished" podID="d33f9d63-34d7-4722-95d5-13bdcb32b99a" containerID="2dae75ee51902fd1a5cb22fba350a24fe1325a6e77293756078944b96466dffe" exitCode=2 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.906946 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9vsj7" event={"ID":"d33f9d63-34d7-4722-95d5-13bdcb32b99a","Type":"ContainerDied","Data":"2dae75ee51902fd1a5cb22fba350a24fe1325a6e77293756078944b96466dffe"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.907023 4823 scope.go:117] "RemoveContainer" containerID="78ffccc041c962174ae10be7c0e04eb2732412f9906f9bc5ed3233dee4fcc42b" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.907680 4823 scope.go:117] "RemoveContainer" containerID="2dae75ee51902fd1a5cb22fba350a24fe1325a6e77293756078944b96466dffe" Mar 20 09:10:21 crc kubenswrapper[4823]: E0320 09:10:21.908045 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-9vsj7_openshift-multus(d33f9d63-34d7-4722-95d5-13bdcb32b99a)\"" pod="openshift-multus/multus-9vsj7" podUID="d33f9d63-34d7-4722-95d5-13bdcb32b99a" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.910648 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovnkube-controller/3.log" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.915812 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovn-acl-logging/0.log" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.916855 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2x54k_1bed1862-5bcd-46fe-bc84-2dbab416086b/ovn-controller/0.log" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.917635 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1" exitCode=0 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.917679 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12" exitCode=0 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.917699 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55" exitCode=0 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.917714 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f" exitCode=0 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.917732 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c" exitCode=0 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.917783 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615" exitCode=0 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.917802 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680" exitCode=143 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.917820 4823 generic.go:334] "Generic (PLEG): container finished" podID="1bed1862-5bcd-46fe-bc84-2dbab416086b" containerID="8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3" exitCode=143 Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.917903 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.917953 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.917981 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918005 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918031 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918054 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918078 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918097 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918111 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918124 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918137 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918152 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918165 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918178 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918190 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918203 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918222 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918243 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918260 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918276 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918289 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918299 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918310 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918321 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918332 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918343 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918354 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918368 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918384 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918399 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918409 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918420 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918430 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918441 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918452 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918462 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918472 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918482 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918497 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" event={"ID":"1bed1862-5bcd-46fe-bc84-2dbab416086b","Type":"ContainerDied","Data":"e1e18bdec402b725c1c518a992d50ffec70ce8b27b41c8c41436593e42fbf51f"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918512 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918524 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918534 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918545 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918555 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918565 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918575 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918585 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918595 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918605 4823 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.918801 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2x54k" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.924174 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" event={"ID":"533a29bc-33f6-4b3c-b222-d601beb87ab8","Type":"ContainerStarted","Data":"ae266d7019a1bd19fb9b757de9ed1b30e474a95df0c927e1b7745d6904357986"} Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.942796 4823 scope.go:117] "RemoveContainer" containerID="9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.982579 4823 scope.go:117] "RemoveContainer" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.991105 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2x54k"] Mar 20 09:10:21 crc kubenswrapper[4823]: I0320 09:10:21.994765 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2x54k"] Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.000243 4823 scope.go:117] "RemoveContainer" containerID="4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.015382 4823 scope.go:117] "RemoveContainer" containerID="2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.061673 4823 scope.go:117] "RemoveContainer" containerID="bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.077587 4823 scope.go:117] "RemoveContainer" containerID="d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.089926 4823 scope.go:117] "RemoveContainer" containerID="46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.101683 4823 scope.go:117] "RemoveContainer" containerID="37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.103393 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bed1862-5bcd-46fe-bc84-2dbab416086b" path="/var/lib/kubelet/pods/1bed1862-5bcd-46fe-bc84-2dbab416086b/volumes" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.113389 4823 scope.go:117] "RemoveContainer" containerID="8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.126917 4823 scope.go:117] "RemoveContainer" containerID="a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.140059 4823 scope.go:117] "RemoveContainer" containerID="9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1" Mar 20 09:10:22 crc kubenswrapper[4823]: E0320 09:10:22.140446 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1\": container with ID starting with 9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1 not found: ID does not exist" containerID="9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.140479 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1"} err="failed to get container status \"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1\": rpc error: code = NotFound desc = could not find container \"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1\": container with ID starting with 9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.140501 4823 scope.go:117] "RemoveContainer" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 09:10:22 crc kubenswrapper[4823]: E0320 09:10:22.140915 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\": container with ID starting with 9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469 not found: ID does not exist" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.140938 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469"} err="failed to get container status \"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\": rpc error: code = NotFound desc = could not find container \"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\": container with ID starting with 9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.140952 4823 scope.go:117] "RemoveContainer" containerID="4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12" Mar 20 09:10:22 crc kubenswrapper[4823]: E0320 09:10:22.141266 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\": container with ID starting with 4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12 not found: ID does not exist" containerID="4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.141286 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12"} err="failed to get container status \"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\": rpc error: code = NotFound desc = could not find container \"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\": container with ID starting with 4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.141300 4823 scope.go:117] "RemoveContainer" containerID="2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55" Mar 20 09:10:22 crc kubenswrapper[4823]: E0320 09:10:22.141644 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\": container with ID starting with 2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55 not found: ID does not exist" containerID="2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.141698 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55"} err="failed to get container status \"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\": rpc error: code = NotFound desc = could not find container \"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\": container with ID starting with 2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.141760 4823 scope.go:117] "RemoveContainer" containerID="bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f" Mar 20 09:10:22 crc kubenswrapper[4823]: E0320 09:10:22.142151 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\": container with ID starting with bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f not found: ID does not exist" containerID="bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.142186 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f"} err="failed to get container status \"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\": rpc error: code = NotFound desc = could not find container \"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\": container with ID starting with bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.142208 4823 scope.go:117] "RemoveContainer" containerID="d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c" Mar 20 09:10:22 crc kubenswrapper[4823]: E0320 09:10:22.142529 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\": container with ID starting with d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c not found: ID does not exist" containerID="d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.142553 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c"} err="failed to get container status \"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\": rpc error: code = NotFound desc = could not find container \"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\": container with ID starting with d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.142568 4823 scope.go:117] "RemoveContainer" containerID="46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615" Mar 20 09:10:22 crc kubenswrapper[4823]: E0320 09:10:22.142902 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\": container with ID starting with 46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615 not found: ID does not exist" containerID="46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.142926 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615"} err="failed to get container status \"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\": rpc error: code = NotFound desc = could not find container \"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\": container with ID starting with 46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.142941 4823 scope.go:117] "RemoveContainer" containerID="37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680" Mar 20 09:10:22 crc kubenswrapper[4823]: E0320 09:10:22.143172 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\": container with ID starting with 37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680 not found: ID does not exist" containerID="37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.143216 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680"} err="failed to get container status \"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\": rpc error: code = NotFound desc = could not find container \"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\": container with ID starting with 37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.143240 4823 scope.go:117] "RemoveContainer" containerID="8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3" Mar 20 09:10:22 crc kubenswrapper[4823]: E0320 09:10:22.143566 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\": container with ID starting with 8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3 not found: ID does not exist" containerID="8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.143590 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3"} err="failed to get container status \"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\": rpc error: code = NotFound desc = could not find container \"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\": container with ID starting with 8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.143607 4823 scope.go:117] "RemoveContainer" containerID="a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48" Mar 20 09:10:22 crc kubenswrapper[4823]: E0320 09:10:22.143984 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\": container with ID starting with a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48 not found: ID does not exist" containerID="a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.144009 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48"} err="failed to get container status \"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\": rpc error: code = NotFound desc = could not find container \"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\": container with ID starting with a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.144024 4823 scope.go:117] "RemoveContainer" containerID="9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.144282 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1"} err="failed to get container status \"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1\": rpc error: code = NotFound desc = could not find container \"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1\": container with ID starting with 9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.144315 4823 scope.go:117] "RemoveContainer" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.144625 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469"} err="failed to get container status \"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\": rpc error: code = NotFound desc = could not find container \"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\": container with ID starting with 9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.144646 4823 scope.go:117] "RemoveContainer" containerID="4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.144886 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12"} err="failed to get container status \"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\": rpc error: code = NotFound desc = could not find container \"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\": container with ID starting with 4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.144919 4823 scope.go:117] "RemoveContainer" containerID="2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.145139 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55"} err="failed to get container status \"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\": rpc error: code = NotFound desc = could not find container \"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\": container with ID starting with 2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.145170 4823 scope.go:117] "RemoveContainer" containerID="bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.145411 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f"} err="failed to get container status \"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\": rpc error: code = NotFound desc = could not find container \"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\": container with ID starting with bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.145436 4823 scope.go:117] "RemoveContainer" containerID="d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.145634 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c"} err="failed to get container status \"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\": rpc error: code = NotFound desc = could not find container \"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\": container with ID starting with d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.145667 4823 scope.go:117] "RemoveContainer" containerID="46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.145978 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615"} err="failed to get container status \"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\": rpc error: code = NotFound desc = could not find container \"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\": container with ID starting with 46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.145998 4823 scope.go:117] "RemoveContainer" containerID="37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.146254 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680"} err="failed to get container status \"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\": rpc error: code = NotFound desc = could not find container \"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\": container with ID starting with 37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.146286 4823 scope.go:117] "RemoveContainer" containerID="8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.146543 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3"} err="failed to get container status \"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\": rpc error: code = NotFound desc = could not find container \"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\": container with ID starting with 8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.146576 4823 scope.go:117] "RemoveContainer" containerID="a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.146857 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48"} err="failed to get container status \"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\": rpc error: code = NotFound desc = could not find container \"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\": container with ID starting with a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.146879 4823 scope.go:117] "RemoveContainer" containerID="9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.147098 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1"} err="failed to get container status \"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1\": rpc error: code = NotFound desc = could not find container \"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1\": container with ID starting with 9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.147117 4823 scope.go:117] "RemoveContainer" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.147329 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469"} err="failed to get container status \"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\": rpc error: code = NotFound desc = could not find container \"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\": container with ID starting with 9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.147352 4823 scope.go:117] "RemoveContainer" containerID="4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.147581 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12"} err="failed to get container status \"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\": rpc error: code = NotFound desc = could not find container \"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\": container with ID starting with 4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.147599 4823 scope.go:117] "RemoveContainer" containerID="2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.147869 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55"} err="failed to get container status \"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\": rpc error: code = NotFound desc = could not find container \"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\": container with ID starting with 2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.147899 4823 scope.go:117] "RemoveContainer" containerID="bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.148134 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f"} err="failed to get container status \"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\": rpc error: code = NotFound desc = could not find container \"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\": container with ID starting with bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.148157 4823 scope.go:117] "RemoveContainer" containerID="d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.148374 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c"} err="failed to get container status \"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\": rpc error: code = NotFound desc = could not find container \"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\": container with ID starting with d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.148400 4823 scope.go:117] "RemoveContainer" containerID="46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.148669 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615"} err="failed to get container status \"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\": rpc error: code = NotFound desc = could not find container \"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\": container with ID starting with 46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.148696 4823 scope.go:117] "RemoveContainer" containerID="37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.148950 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680"} err="failed to get container status \"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\": rpc error: code = NotFound desc = could not find container \"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\": container with ID starting with 37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.148970 4823 scope.go:117] "RemoveContainer" containerID="8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.149192 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3"} err="failed to get container status \"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\": rpc error: code = NotFound desc = could not find container \"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\": container with ID starting with 8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.149209 4823 scope.go:117] "RemoveContainer" containerID="a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.149459 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48"} err="failed to get container status \"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\": rpc error: code = NotFound desc = could not find container \"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\": container with ID starting with a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.149494 4823 scope.go:117] "RemoveContainer" containerID="9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.149786 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1"} err="failed to get container status \"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1\": rpc error: code = NotFound desc = could not find container \"9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1\": container with ID starting with 9749fdb631dbf68792fa1ac3d02e98a5cded2ec53790efec182e74ae025661d1 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.149809 4823 scope.go:117] "RemoveContainer" containerID="9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.150027 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469"} err="failed to get container status \"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\": rpc error: code = NotFound desc = could not find container \"9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469\": container with ID starting with 9b2e434823bc9f12177c779570f95b9401ed9c0f605a8e02d6dc572ed9232469 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.150047 4823 scope.go:117] "RemoveContainer" containerID="4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.150220 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12"} err="failed to get container status \"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\": rpc error: code = NotFound desc = could not find container \"4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12\": container with ID starting with 4ba5e07fc123b14b78e3334e43549c142ab69816613719bf4f0ab7cefde3fc12 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.150238 4823 scope.go:117] "RemoveContainer" containerID="2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.150499 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55"} err="failed to get container status \"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\": rpc error: code = NotFound desc = could not find container \"2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55\": container with ID starting with 2ffc34210d896d3a8dbbd68b09ca2b9b7188e23cd61da2daf91d1bb155a56e55 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.150519 4823 scope.go:117] "RemoveContainer" containerID="bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.150770 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f"} err="failed to get container status \"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\": rpc error: code = NotFound desc = could not find container \"bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f\": container with ID starting with bcbd7f5e3919ba0ab401b432b8187f3698aaa008000781bc754d7098623e3d7f not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.150802 4823 scope.go:117] "RemoveContainer" containerID="d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.150998 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c"} err="failed to get container status \"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\": rpc error: code = NotFound desc = could not find container \"d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c\": container with ID starting with d4ec1bd84d40576ed1cf62c9049ef2a5573ff197e1d0f37087000f2f6cc35b2c not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.151023 4823 scope.go:117] "RemoveContainer" containerID="46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.151261 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615"} err="failed to get container status \"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\": rpc error: code = NotFound desc = could not find container \"46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615\": container with ID starting with 46fcce59919fd2a96e3b0d864db58e8e5c14804ad2f31a465a1db94c6273b615 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.151282 4823 scope.go:117] "RemoveContainer" containerID="37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.151535 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680"} err="failed to get container status \"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\": rpc error: code = NotFound desc = could not find container \"37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680\": container with ID starting with 37ac0120454e0b83c7bcdb22c4c0c594758a75cc09bcb589d4dc7b04ab41a680 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.151611 4823 scope.go:117] "RemoveContainer" containerID="8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.151978 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3"} err="failed to get container status \"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\": rpc error: code = NotFound desc = could not find container \"8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3\": container with ID starting with 8bffeb4b5a123086a6025c2386459104c92696be2972079682e6bc6015cf94c3 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.151997 4823 scope.go:117] "RemoveContainer" containerID="a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.152229 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48"} err="failed to get container status \"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\": rpc error: code = NotFound desc = could not find container \"a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48\": container with ID starting with a70c7a859076469fc1a6fa1d41c3e4b4c09a771e172a25025f6a05addf186b48 not found: ID does not exist" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.932868 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9vsj7_d33f9d63-34d7-4722-95d5-13bdcb32b99a/kube-multus/2.log" Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.935374 4823 generic.go:334] "Generic (PLEG): container finished" podID="533a29bc-33f6-4b3c-b222-d601beb87ab8" containerID="a5e78d480489cd6ba5c07005b3fd194854b53c4d0510d627bd4bc14a18b11e37" exitCode=0 Mar 20 09:10:22 crc kubenswrapper[4823]: I0320 09:10:22.935421 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" event={"ID":"533a29bc-33f6-4b3c-b222-d601beb87ab8","Type":"ContainerDied","Data":"a5e78d480489cd6ba5c07005b3fd194854b53c4d0510d627bd4bc14a18b11e37"} Mar 20 09:10:23 crc kubenswrapper[4823]: I0320 09:10:23.945863 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" event={"ID":"533a29bc-33f6-4b3c-b222-d601beb87ab8","Type":"ContainerStarted","Data":"7d63e50b685531152e55fcf70a763af2084410b73b64f1462afbcf3a13a4cbe9"} Mar 20 09:10:23 crc kubenswrapper[4823]: I0320 09:10:23.946453 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" event={"ID":"533a29bc-33f6-4b3c-b222-d601beb87ab8","Type":"ContainerStarted","Data":"6dc17be902b80a2293a56de1a56e8e272821d7e01164c4b88f126708590a18c9"} Mar 20 09:10:23 crc kubenswrapper[4823]: I0320 09:10:23.946473 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" event={"ID":"533a29bc-33f6-4b3c-b222-d601beb87ab8","Type":"ContainerStarted","Data":"03de7f2a022a15888a6b9fd20286b4f1a42476f22678212d293df044a335f556"} Mar 20 09:10:23 crc kubenswrapper[4823]: I0320 09:10:23.946491 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" event={"ID":"533a29bc-33f6-4b3c-b222-d601beb87ab8","Type":"ContainerStarted","Data":"e127e16b14386b7c7467a67795a15627cf3a4f9ecd4ea5d94d89d629f69dd665"} Mar 20 09:10:23 crc kubenswrapper[4823]: I0320 09:10:23.946509 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" event={"ID":"533a29bc-33f6-4b3c-b222-d601beb87ab8","Type":"ContainerStarted","Data":"bd3698a082c9210f65e41b3444c536f6773f3ae4aa13a84ec59a7b01df66a5a4"} Mar 20 09:10:23 crc kubenswrapper[4823]: I0320 09:10:23.946526 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" event={"ID":"533a29bc-33f6-4b3c-b222-d601beb87ab8","Type":"ContainerStarted","Data":"a80b2399ef0e91d1891f9d9157f1f7d3614a0cd8f923773f3fd2cccb409cfa36"} Mar 20 09:10:25 crc kubenswrapper[4823]: I0320 09:10:25.971209 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" event={"ID":"533a29bc-33f6-4b3c-b222-d601beb87ab8","Type":"ContainerStarted","Data":"ae447d10b74e417c225a0bb6dad9ed27893eb7f4bf40be13f96012752791156d"} Mar 20 09:10:28 crc kubenswrapper[4823]: I0320 09:10:28.995378 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" event={"ID":"533a29bc-33f6-4b3c-b222-d601beb87ab8","Type":"ContainerStarted","Data":"f861254c9a38a17da07d4671565b888a075567a3d5acbe5d85da131e2ab33449"} Mar 20 09:10:28 crc kubenswrapper[4823]: I0320 09:10:28.996181 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:28 crc kubenswrapper[4823]: I0320 09:10:28.996302 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:28 crc kubenswrapper[4823]: I0320 09:10:28.996322 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:29 crc kubenswrapper[4823]: I0320 09:10:29.034635 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:29 crc kubenswrapper[4823]: I0320 09:10:29.041161 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:10:29 crc kubenswrapper[4823]: I0320 09:10:29.044923 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" podStartSLOduration=8.044895026 podStartE2EDuration="8.044895026s" podCreationTimestamp="2026-03-20 09:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:10:29.041043516 +0000 UTC m=+795.726939890" watchObservedRunningTime="2026-03-20 09:10:29.044895026 +0000 UTC m=+795.730791400" Mar 20 09:10:31 crc kubenswrapper[4823]: I0320 09:10:31.494019 4823 scope.go:117] "RemoveContainer" containerID="efc76f0896d85f59ad395378397e44e69e669e1422379d4157219c62a9f1735f" Mar 20 09:10:34 crc kubenswrapper[4823]: I0320 09:10:34.101191 4823 scope.go:117] "RemoveContainer" containerID="2dae75ee51902fd1a5cb22fba350a24fe1325a6e77293756078944b96466dffe" Mar 20 09:10:34 crc kubenswrapper[4823]: E0320 09:10:34.101734 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-9vsj7_openshift-multus(d33f9d63-34d7-4722-95d5-13bdcb32b99a)\"" pod="openshift-multus/multus-9vsj7" podUID="d33f9d63-34d7-4722-95d5-13bdcb32b99a" Mar 20 09:10:39 crc kubenswrapper[4823]: I0320 09:10:39.430010 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:10:39 crc kubenswrapper[4823]: I0320 09:10:39.430501 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:10:48 crc kubenswrapper[4823]: I0320 09:10:48.097352 4823 scope.go:117] "RemoveContainer" containerID="2dae75ee51902fd1a5cb22fba350a24fe1325a6e77293756078944b96466dffe" Mar 20 09:10:48 crc kubenswrapper[4823]: I0320 09:10:48.649521 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9vsj7_d33f9d63-34d7-4722-95d5-13bdcb32b99a/kube-multus/2.log" Mar 20 09:10:48 crc kubenswrapper[4823]: I0320 09:10:48.649836 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9vsj7" event={"ID":"d33f9d63-34d7-4722-95d5-13bdcb32b99a","Type":"ContainerStarted","Data":"f5e83866f7a8a0a7a0c4bd68087335951bdc65533ce8a3ed0f496bade2426bce"} Mar 20 09:10:51 crc kubenswrapper[4823]: I0320 09:10:51.834613 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6qll5" Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.744873 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4"] Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.746520 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.748625 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.753945 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4"] Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.897566 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.897618 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn5fz\" (UniqueName: \"kubernetes.io/projected/66e1071b-3fa5-454b-aca8-ec2fa91d195e-kube-api-access-jn5fz\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.897655 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.998622 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.998673 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn5fz\" (UniqueName: \"kubernetes.io/projected/66e1071b-3fa5-454b-aca8-ec2fa91d195e-kube-api-access-jn5fz\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.998716 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.999339 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:01 crc kubenswrapper[4823]: I0320 09:11:01.999375 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:02 crc kubenswrapper[4823]: I0320 09:11:02.023100 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn5fz\" (UniqueName: \"kubernetes.io/projected/66e1071b-3fa5-454b-aca8-ec2fa91d195e-kube-api-access-jn5fz\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:02 crc kubenswrapper[4823]: I0320 09:11:02.062211 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:02 crc kubenswrapper[4823]: I0320 09:11:02.236658 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4"] Mar 20 09:11:02 crc kubenswrapper[4823]: I0320 09:11:02.923915 4823 generic.go:334] "Generic (PLEG): container finished" podID="66e1071b-3fa5-454b-aca8-ec2fa91d195e" containerID="4987c70977520b168fa6c983ebf49bdecdf2754b2442f46c743e34c7c526c9ea" exitCode=0 Mar 20 09:11:02 crc kubenswrapper[4823]: I0320 09:11:02.923977 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" event={"ID":"66e1071b-3fa5-454b-aca8-ec2fa91d195e","Type":"ContainerDied","Data":"4987c70977520b168fa6c983ebf49bdecdf2754b2442f46c743e34c7c526c9ea"} Mar 20 09:11:02 crc kubenswrapper[4823]: I0320 09:11:02.924334 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" event={"ID":"66e1071b-3fa5-454b-aca8-ec2fa91d195e","Type":"ContainerStarted","Data":"5e229e8ce8272ae8fcef9f08ef0a0273a3b58774961f74ea7e6a8ce1ef535bda"} Mar 20 09:11:02 crc kubenswrapper[4823]: I0320 09:11:02.926060 4823 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.029661 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2dhjv"] Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.031772 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.040106 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2dhjv"] Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.041582 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-utilities\") pod \"redhat-operators-2dhjv\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.042359 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhkdr\" (UniqueName: \"kubernetes.io/projected/7fa1ba59-a823-4576-ab37-53e75f2d11a4-kube-api-access-rhkdr\") pod \"redhat-operators-2dhjv\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.042489 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-catalog-content\") pod \"redhat-operators-2dhjv\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.143622 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhkdr\" (UniqueName: \"kubernetes.io/projected/7fa1ba59-a823-4576-ab37-53e75f2d11a4-kube-api-access-rhkdr\") pod \"redhat-operators-2dhjv\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.143667 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-catalog-content\") pod \"redhat-operators-2dhjv\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.143716 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-utilities\") pod \"redhat-operators-2dhjv\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.144147 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-utilities\") pod \"redhat-operators-2dhjv\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.144633 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-catalog-content\") pod \"redhat-operators-2dhjv\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.180498 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhkdr\" (UniqueName: \"kubernetes.io/projected/7fa1ba59-a823-4576-ab37-53e75f2d11a4-kube-api-access-rhkdr\") pod \"redhat-operators-2dhjv\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.345281 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.549229 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2dhjv"] Mar 20 09:11:04 crc kubenswrapper[4823]: W0320 09:11:04.567701 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fa1ba59_a823_4576_ab37_53e75f2d11a4.slice/crio-7532fdc3fac33ad141cef1dea144d5bb043ec5e958b2b6fe8792eecd58a53905 WatchSource:0}: Error finding container 7532fdc3fac33ad141cef1dea144d5bb043ec5e958b2b6fe8792eecd58a53905: Status 404 returned error can't find the container with id 7532fdc3fac33ad141cef1dea144d5bb043ec5e958b2b6fe8792eecd58a53905 Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.740195 4823 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.938288 4823 generic.go:334] "Generic (PLEG): container finished" podID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerID="d87623967f027236053066ef9d91748550fa93d35dc9c192528902d69e245515" exitCode=0 Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.938368 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2dhjv" event={"ID":"7fa1ba59-a823-4576-ab37-53e75f2d11a4","Type":"ContainerDied","Data":"d87623967f027236053066ef9d91748550fa93d35dc9c192528902d69e245515"} Mar 20 09:11:04 crc kubenswrapper[4823]: I0320 09:11:04.938405 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2dhjv" event={"ID":"7fa1ba59-a823-4576-ab37-53e75f2d11a4","Type":"ContainerStarted","Data":"7532fdc3fac33ad141cef1dea144d5bb043ec5e958b2b6fe8792eecd58a53905"} Mar 20 09:11:05 crc kubenswrapper[4823]: I0320 09:11:05.950017 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2dhjv" event={"ID":"7fa1ba59-a823-4576-ab37-53e75f2d11a4","Type":"ContainerStarted","Data":"18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1"} Mar 20 09:11:06 crc kubenswrapper[4823]: I0320 09:11:06.959916 4823 generic.go:334] "Generic (PLEG): container finished" podID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerID="18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1" exitCode=0 Mar 20 09:11:06 crc kubenswrapper[4823]: I0320 09:11:06.960025 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2dhjv" event={"ID":"7fa1ba59-a823-4576-ab37-53e75f2d11a4","Type":"ContainerDied","Data":"18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1"} Mar 20 09:11:07 crc kubenswrapper[4823]: I0320 09:11:07.969545 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2dhjv" event={"ID":"7fa1ba59-a823-4576-ab37-53e75f2d11a4","Type":"ContainerStarted","Data":"b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f"} Mar 20 09:11:07 crc kubenswrapper[4823]: I0320 09:11:07.997280 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2dhjv" podStartSLOduration=1.391044149 podStartE2EDuration="3.997237897s" podCreationTimestamp="2026-03-20 09:11:04 +0000 UTC" firstStartedPulling="2026-03-20 09:11:04.939837995 +0000 UTC m=+831.625734369" lastFinishedPulling="2026-03-20 09:11:07.546031703 +0000 UTC m=+834.231928117" observedRunningTime="2026-03-20 09:11:07.992733417 +0000 UTC m=+834.678629811" watchObservedRunningTime="2026-03-20 09:11:07.997237897 +0000 UTC m=+834.683134291" Mar 20 09:11:09 crc kubenswrapper[4823]: I0320 09:11:09.430875 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:11:09 crc kubenswrapper[4823]: I0320 09:11:09.430967 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:11:09 crc kubenswrapper[4823]: I0320 09:11:09.431026 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:11:09 crc kubenswrapper[4823]: I0320 09:11:09.431581 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"60ec5c90c519363b5a41f87bdb8567fcf49b737ce1cfeea150ceae7849bf0158"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:11:09 crc kubenswrapper[4823]: I0320 09:11:09.431650 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://60ec5c90c519363b5a41f87bdb8567fcf49b737ce1cfeea150ceae7849bf0158" gracePeriod=600 Mar 20 09:11:09 crc kubenswrapper[4823]: I0320 09:11:09.983008 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="60ec5c90c519363b5a41f87bdb8567fcf49b737ce1cfeea150ceae7849bf0158" exitCode=0 Mar 20 09:11:09 crc kubenswrapper[4823]: I0320 09:11:09.983120 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"60ec5c90c519363b5a41f87bdb8567fcf49b737ce1cfeea150ceae7849bf0158"} Mar 20 09:11:09 crc kubenswrapper[4823]: I0320 09:11:09.983318 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"3c6623b862df4196951c10211c83ea8d06f08a5d6875319545662d0ea4663061"} Mar 20 09:11:09 crc kubenswrapper[4823]: I0320 09:11:09.983341 4823 scope.go:117] "RemoveContainer" containerID="684e65a34e103a8e275908b0f42a447fad127cb196d2c25b351518845765eb92" Mar 20 09:11:14 crc kubenswrapper[4823]: I0320 09:11:14.346413 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:14 crc kubenswrapper[4823]: I0320 09:11:14.347070 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:15 crc kubenswrapper[4823]: I0320 09:11:15.398403 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2dhjv" podUID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerName="registry-server" probeResult="failure" output=< Mar 20 09:11:15 crc kubenswrapper[4823]: timeout: failed to connect service ":50051" within 1s Mar 20 09:11:15 crc kubenswrapper[4823]: > Mar 20 09:11:24 crc kubenswrapper[4823]: I0320 09:11:24.400187 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:24 crc kubenswrapper[4823]: I0320 09:11:24.447567 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:24 crc kubenswrapper[4823]: I0320 09:11:24.630790 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2dhjv"] Mar 20 09:11:26 crc kubenswrapper[4823]: I0320 09:11:26.081909 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2dhjv" podUID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerName="registry-server" containerID="cri-o://b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f" gracePeriod=2 Mar 20 09:11:26 crc kubenswrapper[4823]: I0320 09:11:26.426059 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:26 crc kubenswrapper[4823]: I0320 09:11:26.531616 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhkdr\" (UniqueName: \"kubernetes.io/projected/7fa1ba59-a823-4576-ab37-53e75f2d11a4-kube-api-access-rhkdr\") pod \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " Mar 20 09:11:26 crc kubenswrapper[4823]: I0320 09:11:26.531732 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-catalog-content\") pod \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " Mar 20 09:11:26 crc kubenswrapper[4823]: I0320 09:11:26.531804 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-utilities\") pod \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\" (UID: \"7fa1ba59-a823-4576-ab37-53e75f2d11a4\") " Mar 20 09:11:26 crc kubenswrapper[4823]: I0320 09:11:26.532649 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-utilities" (OuterVolumeSpecName: "utilities") pod "7fa1ba59-a823-4576-ab37-53e75f2d11a4" (UID: "7fa1ba59-a823-4576-ab37-53e75f2d11a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:11:26 crc kubenswrapper[4823]: I0320 09:11:26.538026 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fa1ba59-a823-4576-ab37-53e75f2d11a4-kube-api-access-rhkdr" (OuterVolumeSpecName: "kube-api-access-rhkdr") pod "7fa1ba59-a823-4576-ab37-53e75f2d11a4" (UID: "7fa1ba59-a823-4576-ab37-53e75f2d11a4"). InnerVolumeSpecName "kube-api-access-rhkdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:11:26 crc kubenswrapper[4823]: I0320 09:11:26.633734 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:26 crc kubenswrapper[4823]: I0320 09:11:26.633804 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhkdr\" (UniqueName: \"kubernetes.io/projected/7fa1ba59-a823-4576-ab37-53e75f2d11a4-kube-api-access-rhkdr\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:26 crc kubenswrapper[4823]: I0320 09:11:26.690410 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fa1ba59-a823-4576-ab37-53e75f2d11a4" (UID: "7fa1ba59-a823-4576-ab37-53e75f2d11a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:11:26 crc kubenswrapper[4823]: I0320 09:11:26.734895 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fa1ba59-a823-4576-ab37-53e75f2d11a4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.093471 4823 generic.go:334] "Generic (PLEG): container finished" podID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerID="b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f" exitCode=0 Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.093566 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2dhjv" Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.093558 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2dhjv" event={"ID":"7fa1ba59-a823-4576-ab37-53e75f2d11a4","Type":"ContainerDied","Data":"b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f"} Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.096616 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2dhjv" event={"ID":"7fa1ba59-a823-4576-ab37-53e75f2d11a4","Type":"ContainerDied","Data":"7532fdc3fac33ad141cef1dea144d5bb043ec5e958b2b6fe8792eecd58a53905"} Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.096720 4823 scope.go:117] "RemoveContainer" containerID="b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f" Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.120065 4823 scope.go:117] "RemoveContainer" containerID="18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1" Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.152527 4823 scope.go:117] "RemoveContainer" containerID="d87623967f027236053066ef9d91748550fa93d35dc9c192528902d69e245515" Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.158323 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2dhjv"] Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.165352 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2dhjv"] Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.167009 4823 scope.go:117] "RemoveContainer" containerID="b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f" Mar 20 09:11:27 crc kubenswrapper[4823]: E0320 09:11:27.167586 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f\": container with ID starting with b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f not found: ID does not exist" containerID="b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f" Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.167627 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f"} err="failed to get container status \"b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f\": rpc error: code = NotFound desc = could not find container \"b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f\": container with ID starting with b70dc82829e75a5315711ce667bf1bfd33fb0ddd9cfa5b56563229c68d96f64f not found: ID does not exist" Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.167650 4823 scope.go:117] "RemoveContainer" containerID="18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1" Mar 20 09:11:27 crc kubenswrapper[4823]: E0320 09:11:27.168176 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1\": container with ID starting with 18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1 not found: ID does not exist" containerID="18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1" Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.168206 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1"} err="failed to get container status \"18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1\": rpc error: code = NotFound desc = could not find container \"18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1\": container with ID starting with 18d9173179974be8f441798950170effa68472d0bbfa62273165c212b4f94df1 not found: ID does not exist" Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.168225 4823 scope.go:117] "RemoveContainer" containerID="d87623967f027236053066ef9d91748550fa93d35dc9c192528902d69e245515" Mar 20 09:11:27 crc kubenswrapper[4823]: E0320 09:11:27.168505 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d87623967f027236053066ef9d91748550fa93d35dc9c192528902d69e245515\": container with ID starting with d87623967f027236053066ef9d91748550fa93d35dc9c192528902d69e245515 not found: ID does not exist" containerID="d87623967f027236053066ef9d91748550fa93d35dc9c192528902d69e245515" Mar 20 09:11:27 crc kubenswrapper[4823]: I0320 09:11:27.168534 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d87623967f027236053066ef9d91748550fa93d35dc9c192528902d69e245515"} err="failed to get container status \"d87623967f027236053066ef9d91748550fa93d35dc9c192528902d69e245515\": rpc error: code = NotFound desc = could not find container \"d87623967f027236053066ef9d91748550fa93d35dc9c192528902d69e245515\": container with ID starting with d87623967f027236053066ef9d91748550fa93d35dc9c192528902d69e245515 not found: ID does not exist" Mar 20 09:11:28 crc kubenswrapper[4823]: I0320 09:11:28.102904 4823 generic.go:334] "Generic (PLEG): container finished" podID="66e1071b-3fa5-454b-aca8-ec2fa91d195e" containerID="e31ff731cae365740adb71d05df4ec9c3e0dea63d31f8167cab2fb9a6f775569" exitCode=0 Mar 20 09:11:28 crc kubenswrapper[4823]: I0320 09:11:28.109730 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" path="/var/lib/kubelet/pods/7fa1ba59-a823-4576-ab37-53e75f2d11a4/volumes" Mar 20 09:11:28 crc kubenswrapper[4823]: I0320 09:11:28.111257 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" event={"ID":"66e1071b-3fa5-454b-aca8-ec2fa91d195e","Type":"ContainerDied","Data":"e31ff731cae365740adb71d05df4ec9c3e0dea63d31f8167cab2fb9a6f775569"} Mar 20 09:11:29 crc kubenswrapper[4823]: I0320 09:11:29.116074 4823 generic.go:334] "Generic (PLEG): container finished" podID="66e1071b-3fa5-454b-aca8-ec2fa91d195e" containerID="9a77a2dfd24be9270d1e1e0625854d88fe25f721eb89366d41a0a5220037df41" exitCode=0 Mar 20 09:11:29 crc kubenswrapper[4823]: I0320 09:11:29.116187 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" event={"ID":"66e1071b-3fa5-454b-aca8-ec2fa91d195e","Type":"ContainerDied","Data":"9a77a2dfd24be9270d1e1e0625854d88fe25f721eb89366d41a0a5220037df41"} Mar 20 09:11:30 crc kubenswrapper[4823]: I0320 09:11:30.390918 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:30 crc kubenswrapper[4823]: I0320 09:11:30.580602 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-bundle\") pod \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " Mar 20 09:11:30 crc kubenswrapper[4823]: I0320 09:11:30.580847 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn5fz\" (UniqueName: \"kubernetes.io/projected/66e1071b-3fa5-454b-aca8-ec2fa91d195e-kube-api-access-jn5fz\") pod \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " Mar 20 09:11:30 crc kubenswrapper[4823]: I0320 09:11:30.580895 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-util\") pod \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\" (UID: \"66e1071b-3fa5-454b-aca8-ec2fa91d195e\") " Mar 20 09:11:30 crc kubenswrapper[4823]: I0320 09:11:30.582006 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-bundle" (OuterVolumeSpecName: "bundle") pod "66e1071b-3fa5-454b-aca8-ec2fa91d195e" (UID: "66e1071b-3fa5-454b-aca8-ec2fa91d195e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:11:30 crc kubenswrapper[4823]: I0320 09:11:30.590098 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66e1071b-3fa5-454b-aca8-ec2fa91d195e-kube-api-access-jn5fz" (OuterVolumeSpecName: "kube-api-access-jn5fz") pod "66e1071b-3fa5-454b-aca8-ec2fa91d195e" (UID: "66e1071b-3fa5-454b-aca8-ec2fa91d195e"). InnerVolumeSpecName "kube-api-access-jn5fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:11:30 crc kubenswrapper[4823]: I0320 09:11:30.608318 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-util" (OuterVolumeSpecName: "util") pod "66e1071b-3fa5-454b-aca8-ec2fa91d195e" (UID: "66e1071b-3fa5-454b-aca8-ec2fa91d195e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:11:30 crc kubenswrapper[4823]: I0320 09:11:30.681946 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn5fz\" (UniqueName: \"kubernetes.io/projected/66e1071b-3fa5-454b-aca8-ec2fa91d195e-kube-api-access-jn5fz\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:30 crc kubenswrapper[4823]: I0320 09:11:30.681982 4823 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-util\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:30 crc kubenswrapper[4823]: I0320 09:11:30.681993 4823 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66e1071b-3fa5-454b-aca8-ec2fa91d195e-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:31 crc kubenswrapper[4823]: I0320 09:11:31.128472 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" event={"ID":"66e1071b-3fa5-454b-aca8-ec2fa91d195e","Type":"ContainerDied","Data":"5e229e8ce8272ae8fcef9f08ef0a0273a3b58774961f74ea7e6a8ce1ef535bda"} Mar 20 09:11:31 crc kubenswrapper[4823]: I0320 09:11:31.128701 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e229e8ce8272ae8fcef9f08ef0a0273a3b58774961f74ea7e6a8ce1ef535bda" Mar 20 09:11:31 crc kubenswrapper[4823]: I0320 09:11:31.128554 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.503431 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-mfn9c"] Mar 20 09:11:33 crc kubenswrapper[4823]: E0320 09:11:33.504035 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerName="registry-server" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.504047 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerName="registry-server" Mar 20 09:11:33 crc kubenswrapper[4823]: E0320 09:11:33.504060 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66e1071b-3fa5-454b-aca8-ec2fa91d195e" containerName="extract" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.504066 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="66e1071b-3fa5-454b-aca8-ec2fa91d195e" containerName="extract" Mar 20 09:11:33 crc kubenswrapper[4823]: E0320 09:11:33.504078 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66e1071b-3fa5-454b-aca8-ec2fa91d195e" containerName="util" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.504084 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="66e1071b-3fa5-454b-aca8-ec2fa91d195e" containerName="util" Mar 20 09:11:33 crc kubenswrapper[4823]: E0320 09:11:33.504092 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66e1071b-3fa5-454b-aca8-ec2fa91d195e" containerName="pull" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.504098 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="66e1071b-3fa5-454b-aca8-ec2fa91d195e" containerName="pull" Mar 20 09:11:33 crc kubenswrapper[4823]: E0320 09:11:33.504105 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerName="extract-content" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.504111 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerName="extract-content" Mar 20 09:11:33 crc kubenswrapper[4823]: E0320 09:11:33.504119 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerName="extract-utilities" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.504126 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerName="extract-utilities" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.504212 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fa1ba59-a823-4576-ab37-53e75f2d11a4" containerName="registry-server" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.504226 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="66e1071b-3fa5-454b-aca8-ec2fa91d195e" containerName="extract" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.504565 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-mfn9c" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.506375 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.506928 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.508276 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-9nvgn" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.523697 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-mfn9c"] Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.618054 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htfgv\" (UniqueName: \"kubernetes.io/projected/b5f0a310-2e6a-4ed5-8703-032996455052-kube-api-access-htfgv\") pod \"nmstate-operator-796d4cfff4-mfn9c\" (UID: \"b5f0a310-2e6a-4ed5-8703-032996455052\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-mfn9c" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.718775 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htfgv\" (UniqueName: \"kubernetes.io/projected/b5f0a310-2e6a-4ed5-8703-032996455052-kube-api-access-htfgv\") pod \"nmstate-operator-796d4cfff4-mfn9c\" (UID: \"b5f0a310-2e6a-4ed5-8703-032996455052\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-mfn9c" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.735831 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htfgv\" (UniqueName: \"kubernetes.io/projected/b5f0a310-2e6a-4ed5-8703-032996455052-kube-api-access-htfgv\") pod \"nmstate-operator-796d4cfff4-mfn9c\" (UID: \"b5f0a310-2e6a-4ed5-8703-032996455052\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-mfn9c" Mar 20 09:11:33 crc kubenswrapper[4823]: I0320 09:11:33.825792 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-mfn9c" Mar 20 09:11:34 crc kubenswrapper[4823]: I0320 09:11:34.257227 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-mfn9c"] Mar 20 09:11:35 crc kubenswrapper[4823]: I0320 09:11:35.159053 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-mfn9c" event={"ID":"b5f0a310-2e6a-4ed5-8703-032996455052","Type":"ContainerStarted","Data":"bf1a6c05fe956623d46c6c55d7b7271cc5f9b6cfcb7a6376769f1b7988246adb"} Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.171509 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-mfn9c" event={"ID":"b5f0a310-2e6a-4ed5-8703-032996455052","Type":"ContainerStarted","Data":"ba165b7496ace0d1c70ced391913d137a6dca0ede528acafe3bf2978f1476a03"} Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.677063 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-mfn9c" podStartSLOduration=2.5734094560000003 podStartE2EDuration="4.677033251s" podCreationTimestamp="2026-03-20 09:11:33 +0000 UTC" firstStartedPulling="2026-03-20 09:11:34.264819791 +0000 UTC m=+860.950716165" lastFinishedPulling="2026-03-20 09:11:36.368443586 +0000 UTC m=+863.054339960" observedRunningTime="2026-03-20 09:11:37.192581243 +0000 UTC m=+863.878477657" watchObservedRunningTime="2026-03-20 09:11:37.677033251 +0000 UTC m=+864.362929655" Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.681810 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hbn4z"] Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.683691 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.696349 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hbn4z"] Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.874729 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-utilities\") pod \"certified-operators-hbn4z\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.874968 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-catalog-content\") pod \"certified-operators-hbn4z\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.875080 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtl45\" (UniqueName: \"kubernetes.io/projected/b6998581-faf4-4ab5-bc20-e1f0467f881d-kube-api-access-qtl45\") pod \"certified-operators-hbn4z\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.976398 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-utilities\") pod \"certified-operators-hbn4z\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.976517 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-catalog-content\") pod \"certified-operators-hbn4z\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.976567 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtl45\" (UniqueName: \"kubernetes.io/projected/b6998581-faf4-4ab5-bc20-e1f0467f881d-kube-api-access-qtl45\") pod \"certified-operators-hbn4z\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.976938 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-utilities\") pod \"certified-operators-hbn4z\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.977039 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-catalog-content\") pod \"certified-operators-hbn4z\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:37 crc kubenswrapper[4823]: I0320 09:11:37.994638 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtl45\" (UniqueName: \"kubernetes.io/projected/b6998581-faf4-4ab5-bc20-e1f0467f881d-kube-api-access-qtl45\") pod \"certified-operators-hbn4z\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:38 crc kubenswrapper[4823]: I0320 09:11:38.003297 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:38 crc kubenswrapper[4823]: I0320 09:11:38.241849 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hbn4z"] Mar 20 09:11:39 crc kubenswrapper[4823]: I0320 09:11:39.192006 4823 generic.go:334] "Generic (PLEG): container finished" podID="b6998581-faf4-4ab5-bc20-e1f0467f881d" containerID="eb05340d441f502026f1a83628d5230732cd88f064226103d7685d668408de36" exitCode=0 Mar 20 09:11:39 crc kubenswrapper[4823]: I0320 09:11:39.192376 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbn4z" event={"ID":"b6998581-faf4-4ab5-bc20-e1f0467f881d","Type":"ContainerDied","Data":"eb05340d441f502026f1a83628d5230732cd88f064226103d7685d668408de36"} Mar 20 09:11:39 crc kubenswrapper[4823]: I0320 09:11:39.192420 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbn4z" event={"ID":"b6998581-faf4-4ab5-bc20-e1f0467f881d","Type":"ContainerStarted","Data":"f9485eea221f72fa1fb9a8a45e133a294243e7ed932a19922352932f1cff380f"} Mar 20 09:11:41 crc kubenswrapper[4823]: I0320 09:11:41.205647 4823 generic.go:334] "Generic (PLEG): container finished" podID="b6998581-faf4-4ab5-bc20-e1f0467f881d" containerID="f59365e42f4b2911a643e15c142b70cf04c844ac6f34d378dcc2fa6b002065f4" exitCode=0 Mar 20 09:11:41 crc kubenswrapper[4823]: I0320 09:11:41.205733 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbn4z" event={"ID":"b6998581-faf4-4ab5-bc20-e1f0467f881d","Type":"ContainerDied","Data":"f59365e42f4b2911a643e15c142b70cf04c844ac6f34d378dcc2fa6b002065f4"} Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.213907 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbn4z" event={"ID":"b6998581-faf4-4ab5-bc20-e1f0467f881d","Type":"ContainerStarted","Data":"c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658"} Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.239225 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hbn4z" podStartSLOduration=2.747566369 podStartE2EDuration="5.239200174s" podCreationTimestamp="2026-03-20 09:11:37 +0000 UTC" firstStartedPulling="2026-03-20 09:11:39.194775257 +0000 UTC m=+865.880671661" lastFinishedPulling="2026-03-20 09:11:41.686409052 +0000 UTC m=+868.372305466" observedRunningTime="2026-03-20 09:11:42.233343542 +0000 UTC m=+868.919239926" watchObservedRunningTime="2026-03-20 09:11:42.239200174 +0000 UTC m=+868.925096568" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.281599 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw"] Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.282615 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.284698 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-rftdz" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.288190 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw"] Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.312323 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5"] Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.313484 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.317301 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.323610 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5"] Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.333966 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwls9\" (UniqueName: \"kubernetes.io/projected/8af45186-c303-4a66-ad3a-bd1d59df752e-kube-api-access-nwls9\") pod \"nmstate-metrics-9b8c8685d-lm2pw\" (UID: \"8af45186-c303-4a66-ad3a-bd1d59df752e\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.334012 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lrxn\" (UniqueName: \"kubernetes.io/projected/a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e-kube-api-access-4lrxn\") pod \"nmstate-webhook-5f558f5558-qfwd5\" (UID: \"a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.334084 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-qfwd5\" (UID: \"a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.346444 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-rm282"] Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.347299 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.434973 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4"] Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.435099 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-nmstate-lock\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.435146 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwls9\" (UniqueName: \"kubernetes.io/projected/8af45186-c303-4a66-ad3a-bd1d59df752e-kube-api-access-nwls9\") pod \"nmstate-metrics-9b8c8685d-lm2pw\" (UID: \"8af45186-c303-4a66-ad3a-bd1d59df752e\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.435170 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lrxn\" (UniqueName: \"kubernetes.io/projected/a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e-kube-api-access-4lrxn\") pod \"nmstate-webhook-5f558f5558-qfwd5\" (UID: \"a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.435281 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-qfwd5\" (UID: \"a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.435342 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-ovs-socket\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.435377 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9mjn\" (UniqueName: \"kubernetes.io/projected/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-kube-api-access-w9mjn\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.435416 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-dbus-socket\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: E0320 09:11:42.435612 4823 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Mar 20 09:11:42 crc kubenswrapper[4823]: E0320 09:11:42.435660 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e-tls-key-pair podName:a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e nodeName:}" failed. No retries permitted until 2026-03-20 09:11:42.93564382 +0000 UTC m=+869.621540194 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e-tls-key-pair") pod "nmstate-webhook-5f558f5558-qfwd5" (UID: "a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e") : secret "openshift-nmstate-webhook" not found Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.435825 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.460638 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.461350 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-x7c46" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.480862 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.485004 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4"] Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.495863 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwls9\" (UniqueName: \"kubernetes.io/projected/8af45186-c303-4a66-ad3a-bd1d59df752e-kube-api-access-nwls9\") pod \"nmstate-metrics-9b8c8685d-lm2pw\" (UID: \"8af45186-c303-4a66-ad3a-bd1d59df752e\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.498562 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lrxn\" (UniqueName: \"kubernetes.io/projected/a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e-kube-api-access-4lrxn\") pod \"nmstate-webhook-5f558f5558-qfwd5\" (UID: \"a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.544295 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-ovs-socket\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.544351 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/23d06143-924e-49a4-bfcd-d1e29cbd247a-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-58bd4\" (UID: \"23d06143-924e-49a4-bfcd-d1e29cbd247a\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.544381 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9mjn\" (UniqueName: \"kubernetes.io/projected/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-kube-api-access-w9mjn\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.544412 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-dbus-socket\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.544442 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdr8v\" (UniqueName: \"kubernetes.io/projected/23d06143-924e-49a4-bfcd-d1e29cbd247a-kube-api-access-bdr8v\") pod \"nmstate-console-plugin-86f58fcf4-58bd4\" (UID: \"23d06143-924e-49a4-bfcd-d1e29cbd247a\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.544486 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/23d06143-924e-49a4-bfcd-d1e29cbd247a-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-58bd4\" (UID: \"23d06143-924e-49a4-bfcd-d1e29cbd247a\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.544509 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-nmstate-lock\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.544596 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-nmstate-lock\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.544644 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-ovs-socket\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.545096 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-dbus-socket\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.565444 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9mjn\" (UniqueName: \"kubernetes.io/projected/76f878ff-14b3-45f2-9c2b-6c4995ab52b2-kube-api-access-w9mjn\") pod \"nmstate-handler-rm282\" (UID: \"76f878ff-14b3-45f2-9c2b-6c4995ab52b2\") " pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.598425 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.645093 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdr8v\" (UniqueName: \"kubernetes.io/projected/23d06143-924e-49a4-bfcd-d1e29cbd247a-kube-api-access-bdr8v\") pod \"nmstate-console-plugin-86f58fcf4-58bd4\" (UID: \"23d06143-924e-49a4-bfcd-d1e29cbd247a\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.645382 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/23d06143-924e-49a4-bfcd-d1e29cbd247a-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-58bd4\" (UID: \"23d06143-924e-49a4-bfcd-d1e29cbd247a\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.645441 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/23d06143-924e-49a4-bfcd-d1e29cbd247a-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-58bd4\" (UID: \"23d06143-924e-49a4-bfcd-d1e29cbd247a\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:42 crc kubenswrapper[4823]: E0320 09:11:42.645554 4823 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 20 09:11:42 crc kubenswrapper[4823]: E0320 09:11:42.645599 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/23d06143-924e-49a4-bfcd-d1e29cbd247a-plugin-serving-cert podName:23d06143-924e-49a4-bfcd-d1e29cbd247a nodeName:}" failed. No retries permitted until 2026-03-20 09:11:43.145585756 +0000 UTC m=+869.831482130 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/23d06143-924e-49a4-bfcd-d1e29cbd247a-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-58bd4" (UID: "23d06143-924e-49a4-bfcd-d1e29cbd247a") : secret "plugin-serving-cert" not found Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.646354 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/23d06143-924e-49a4-bfcd-d1e29cbd247a-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-58bd4\" (UID: \"23d06143-924e-49a4-bfcd-d1e29cbd247a\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.660298 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.673339 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-58d6f8d54f-znrsb"] Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.674248 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.677684 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdr8v\" (UniqueName: \"kubernetes.io/projected/23d06143-924e-49a4-bfcd-d1e29cbd247a-kube-api-access-bdr8v\") pod \"nmstate-console-plugin-86f58fcf4-58bd4\" (UID: \"23d06143-924e-49a4-bfcd-d1e29cbd247a\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.685151 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-58d6f8d54f-znrsb"] Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.746209 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-trusted-ca-bundle\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.746264 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-console-config\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.746298 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-oauth-serving-cert\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.746329 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e47c716b-40a5-49c5-86d3-29998224ad57-console-oauth-config\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.746363 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hngh5\" (UniqueName: \"kubernetes.io/projected/e47c716b-40a5-49c5-86d3-29998224ad57-kube-api-access-hngh5\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.746380 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-service-ca\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.746398 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e47c716b-40a5-49c5-86d3-29998224ad57-console-serving-cert\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.838304 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw"] Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.847552 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hngh5\" (UniqueName: \"kubernetes.io/projected/e47c716b-40a5-49c5-86d3-29998224ad57-kube-api-access-hngh5\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.847596 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-service-ca\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.847624 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e47c716b-40a5-49c5-86d3-29998224ad57-console-serving-cert\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.847652 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-trusted-ca-bundle\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.847690 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-console-config\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.847707 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-oauth-serving-cert\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.847746 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e47c716b-40a5-49c5-86d3-29998224ad57-console-oauth-config\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.849240 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-console-config\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.849242 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-service-ca\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.849433 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-trusted-ca-bundle\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.849889 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e47c716b-40a5-49c5-86d3-29998224ad57-oauth-serving-cert\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.851871 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e47c716b-40a5-49c5-86d3-29998224ad57-console-serving-cert\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.852365 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e47c716b-40a5-49c5-86d3-29998224ad57-console-oauth-config\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.862406 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hngh5\" (UniqueName: \"kubernetes.io/projected/e47c716b-40a5-49c5-86d3-29998224ad57-kube-api-access-hngh5\") pod \"console-58d6f8d54f-znrsb\" (UID: \"e47c716b-40a5-49c5-86d3-29998224ad57\") " pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.949197 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-qfwd5\" (UID: \"a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.952043 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-qfwd5\" (UID: \"a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" Mar 20 09:11:42 crc kubenswrapper[4823]: I0320 09:11:42.995230 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:43 crc kubenswrapper[4823]: I0320 09:11:43.150540 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/23d06143-924e-49a4-bfcd-d1e29cbd247a-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-58bd4\" (UID: \"23d06143-924e-49a4-bfcd-d1e29cbd247a\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:43 crc kubenswrapper[4823]: I0320 09:11:43.154440 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/23d06143-924e-49a4-bfcd-d1e29cbd247a-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-58bd4\" (UID: \"23d06143-924e-49a4-bfcd-d1e29cbd247a\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:43 crc kubenswrapper[4823]: I0320 09:11:43.189744 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-58d6f8d54f-znrsb"] Mar 20 09:11:43 crc kubenswrapper[4823]: W0320 09:11:43.197112 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode47c716b_40a5_49c5_86d3_29998224ad57.slice/crio-2a09434f558b6476ae8ad84a345a0535b1e7146487083e53dd23e9019f018a53 WatchSource:0}: Error finding container 2a09434f558b6476ae8ad84a345a0535b1e7146487083e53dd23e9019f018a53: Status 404 returned error can't find the container with id 2a09434f558b6476ae8ad84a345a0535b1e7146487083e53dd23e9019f018a53 Mar 20 09:11:43 crc kubenswrapper[4823]: I0320 09:11:43.223265 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-58d6f8d54f-znrsb" event={"ID":"e47c716b-40a5-49c5-86d3-29998224ad57","Type":"ContainerStarted","Data":"2a09434f558b6476ae8ad84a345a0535b1e7146487083e53dd23e9019f018a53"} Mar 20 09:11:43 crc kubenswrapper[4823]: I0320 09:11:43.224036 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw" event={"ID":"8af45186-c303-4a66-ad3a-bd1d59df752e","Type":"ContainerStarted","Data":"d4239268e7d80f4108b09b0de66a78f4090ad713c869a2bd61ade6ae8d3a5dd1"} Mar 20 09:11:43 crc kubenswrapper[4823]: I0320 09:11:43.225817 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-rm282" event={"ID":"76f878ff-14b3-45f2-9c2b-6c4995ab52b2","Type":"ContainerStarted","Data":"f998070da4c454efdfb22b70190fd0be41eb29dbc085978fc95e6f444575bb4d"} Mar 20 09:11:43 crc kubenswrapper[4823]: I0320 09:11:43.226060 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" Mar 20 09:11:43 crc kubenswrapper[4823]: I0320 09:11:43.353692 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" Mar 20 09:11:43 crc kubenswrapper[4823]: I0320 09:11:43.432432 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5"] Mar 20 09:11:43 crc kubenswrapper[4823]: I0320 09:11:43.558910 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4"] Mar 20 09:11:43 crc kubenswrapper[4823]: W0320 09:11:43.564322 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23d06143_924e_49a4_bfcd_d1e29cbd247a.slice/crio-5d00067df00853e0e465ad4a3aa31a9ddf4dc46f82d6c35b72f014aab86c7c3c WatchSource:0}: Error finding container 5d00067df00853e0e465ad4a3aa31a9ddf4dc46f82d6c35b72f014aab86c7c3c: Status 404 returned error can't find the container with id 5d00067df00853e0e465ad4a3aa31a9ddf4dc46f82d6c35b72f014aab86c7c3c Mar 20 09:11:44 crc kubenswrapper[4823]: I0320 09:11:44.239793 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" event={"ID":"23d06143-924e-49a4-bfcd-d1e29cbd247a","Type":"ContainerStarted","Data":"5d00067df00853e0e465ad4a3aa31a9ddf4dc46f82d6c35b72f014aab86c7c3c"} Mar 20 09:11:44 crc kubenswrapper[4823]: I0320 09:11:44.241512 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" event={"ID":"a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e","Type":"ContainerStarted","Data":"18583fb1b275ddda21cc1e61946586805ec23351f1d7d8ce182acedec2a76820"} Mar 20 09:11:44 crc kubenswrapper[4823]: I0320 09:11:44.242999 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-58d6f8d54f-znrsb" event={"ID":"e47c716b-40a5-49c5-86d3-29998224ad57","Type":"ContainerStarted","Data":"497a287d352a3683dc46341c309538f07203406d3718ac32987127d89a1b9847"} Mar 20 09:11:44 crc kubenswrapper[4823]: I0320 09:11:44.268342 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-58d6f8d54f-znrsb" podStartSLOduration=2.268320484 podStartE2EDuration="2.268320484s" podCreationTimestamp="2026-03-20 09:11:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:11:44.267298413 +0000 UTC m=+870.953194797" watchObservedRunningTime="2026-03-20 09:11:44.268320484 +0000 UTC m=+870.954216858" Mar 20 09:11:46 crc kubenswrapper[4823]: I0320 09:11:46.256828 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw" event={"ID":"8af45186-c303-4a66-ad3a-bd1d59df752e","Type":"ContainerStarted","Data":"c421c966d54d54f2bd1bf9e9a7a0b970777d5e2eb1faa68a721101b824148c5b"} Mar 20 09:11:46 crc kubenswrapper[4823]: I0320 09:11:46.258561 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" event={"ID":"a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e","Type":"ContainerStarted","Data":"de940985ea674ac615c99f522090984040f2ea6637a0e5bfec3a189d5d975d89"} Mar 20 09:11:46 crc kubenswrapper[4823]: I0320 09:11:46.259345 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" Mar 20 09:11:46 crc kubenswrapper[4823]: I0320 09:11:46.260545 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-rm282" event={"ID":"76f878ff-14b3-45f2-9c2b-6c4995ab52b2","Type":"ContainerStarted","Data":"b6f3f33979d5c4f666b9a7c37113b8d2f11f5c2b44dfbd2c08beb2422f948cf8"} Mar 20 09:11:46 crc kubenswrapper[4823]: I0320 09:11:46.261087 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:46 crc kubenswrapper[4823]: I0320 09:11:46.295164 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" podStartSLOduration=2.471404306 podStartE2EDuration="4.295144843s" podCreationTimestamp="2026-03-20 09:11:42 +0000 UTC" firstStartedPulling="2026-03-20 09:11:43.443469166 +0000 UTC m=+870.129365540" lastFinishedPulling="2026-03-20 09:11:45.267209703 +0000 UTC m=+871.953106077" observedRunningTime="2026-03-20 09:11:46.274897894 +0000 UTC m=+872.960794268" watchObservedRunningTime="2026-03-20 09:11:46.295144843 +0000 UTC m=+872.981041217" Mar 20 09:11:46 crc kubenswrapper[4823]: I0320 09:11:46.299021 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-rm282" podStartSLOduration=1.747632801 podStartE2EDuration="4.299010784s" podCreationTimestamp="2026-03-20 09:11:42 +0000 UTC" firstStartedPulling="2026-03-20 09:11:42.70008436 +0000 UTC m=+869.385980734" lastFinishedPulling="2026-03-20 09:11:45.251462343 +0000 UTC m=+871.937358717" observedRunningTime="2026-03-20 09:11:46.290434317 +0000 UTC m=+872.976330691" watchObservedRunningTime="2026-03-20 09:11:46.299010784 +0000 UTC m=+872.984907158" Mar 20 09:11:47 crc kubenswrapper[4823]: I0320 09:11:47.268130 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" event={"ID":"23d06143-924e-49a4-bfcd-d1e29cbd247a","Type":"ContainerStarted","Data":"c4cd6ecc662aa2cb86fa2df84eba663b96e5ccc7aa31525f2cd548d6180c92ec"} Mar 20 09:11:47 crc kubenswrapper[4823]: I0320 09:11:47.284323 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58bd4" podStartSLOduration=2.533359703 podStartE2EDuration="5.284307459s" podCreationTimestamp="2026-03-20 09:11:42 +0000 UTC" firstStartedPulling="2026-03-20 09:11:43.566674456 +0000 UTC m=+870.252570830" lastFinishedPulling="2026-03-20 09:11:46.317622212 +0000 UTC m=+873.003518586" observedRunningTime="2026-03-20 09:11:47.28144623 +0000 UTC m=+873.967342624" watchObservedRunningTime="2026-03-20 09:11:47.284307459 +0000 UTC m=+873.970203833" Mar 20 09:11:48 crc kubenswrapper[4823]: I0320 09:11:48.004018 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:48 crc kubenswrapper[4823]: I0320 09:11:48.004098 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:48 crc kubenswrapper[4823]: I0320 09:11:48.063557 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:48 crc kubenswrapper[4823]: I0320 09:11:48.279632 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw" event={"ID":"8af45186-c303-4a66-ad3a-bd1d59df752e","Type":"ContainerStarted","Data":"72f17cc4fd514437729734064aa863c4f228a26be52a281a330719067dc2ca76"} Mar 20 09:11:48 crc kubenswrapper[4823]: I0320 09:11:48.298484 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lm2pw" podStartSLOduration=1.403784153 podStartE2EDuration="6.298456662s" podCreationTimestamp="2026-03-20 09:11:42 +0000 UTC" firstStartedPulling="2026-03-20 09:11:42.845318984 +0000 UTC m=+869.531215358" lastFinishedPulling="2026-03-20 09:11:47.739991493 +0000 UTC m=+874.425887867" observedRunningTime="2026-03-20 09:11:48.29517089 +0000 UTC m=+874.981067304" watchObservedRunningTime="2026-03-20 09:11:48.298456662 +0000 UTC m=+874.984353076" Mar 20 09:11:48 crc kubenswrapper[4823]: I0320 09:11:48.328504 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:50 crc kubenswrapper[4823]: I0320 09:11:50.468853 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hbn4z"] Mar 20 09:11:50 crc kubenswrapper[4823]: I0320 09:11:50.469387 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hbn4z" podUID="b6998581-faf4-4ab5-bc20-e1f0467f881d" containerName="registry-server" containerID="cri-o://c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658" gracePeriod=2 Mar 20 09:11:50 crc kubenswrapper[4823]: I0320 09:11:50.881834 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:50 crc kubenswrapper[4823]: I0320 09:11:50.991155 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtl45\" (UniqueName: \"kubernetes.io/projected/b6998581-faf4-4ab5-bc20-e1f0467f881d-kube-api-access-qtl45\") pod \"b6998581-faf4-4ab5-bc20-e1f0467f881d\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " Mar 20 09:11:50 crc kubenswrapper[4823]: I0320 09:11:50.991257 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-utilities\") pod \"b6998581-faf4-4ab5-bc20-e1f0467f881d\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " Mar 20 09:11:50 crc kubenswrapper[4823]: I0320 09:11:50.991286 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-catalog-content\") pod \"b6998581-faf4-4ab5-bc20-e1f0467f881d\" (UID: \"b6998581-faf4-4ab5-bc20-e1f0467f881d\") " Mar 20 09:11:50 crc kubenswrapper[4823]: I0320 09:11:50.992861 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-utilities" (OuterVolumeSpecName: "utilities") pod "b6998581-faf4-4ab5-bc20-e1f0467f881d" (UID: "b6998581-faf4-4ab5-bc20-e1f0467f881d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:11:50 crc kubenswrapper[4823]: I0320 09:11:50.999210 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6998581-faf4-4ab5-bc20-e1f0467f881d-kube-api-access-qtl45" (OuterVolumeSpecName: "kube-api-access-qtl45") pod "b6998581-faf4-4ab5-bc20-e1f0467f881d" (UID: "b6998581-faf4-4ab5-bc20-e1f0467f881d"). InnerVolumeSpecName "kube-api-access-qtl45". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.092852 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtl45\" (UniqueName: \"kubernetes.io/projected/b6998581-faf4-4ab5-bc20-e1f0467f881d-kube-api-access-qtl45\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.092910 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.299502 4823 generic.go:334] "Generic (PLEG): container finished" podID="b6998581-faf4-4ab5-bc20-e1f0467f881d" containerID="c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658" exitCode=0 Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.299595 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbn4z" event={"ID":"b6998581-faf4-4ab5-bc20-e1f0467f881d","Type":"ContainerDied","Data":"c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658"} Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.299871 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbn4z" event={"ID":"b6998581-faf4-4ab5-bc20-e1f0467f881d","Type":"ContainerDied","Data":"f9485eea221f72fa1fb9a8a45e133a294243e7ed932a19922352932f1cff380f"} Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.299677 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbn4z" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.299943 4823 scope.go:117] "RemoveContainer" containerID="c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.323726 4823 scope.go:117] "RemoveContainer" containerID="f59365e42f4b2911a643e15c142b70cf04c844ac6f34d378dcc2fa6b002065f4" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.346586 4823 scope.go:117] "RemoveContainer" containerID="eb05340d441f502026f1a83628d5230732cd88f064226103d7685d668408de36" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.364834 4823 scope.go:117] "RemoveContainer" containerID="c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658" Mar 20 09:11:51 crc kubenswrapper[4823]: E0320 09:11:51.365525 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658\": container with ID starting with c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658 not found: ID does not exist" containerID="c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.365601 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658"} err="failed to get container status \"c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658\": rpc error: code = NotFound desc = could not find container \"c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658\": container with ID starting with c567350446b8341f20936185d542614109fbb310a7e428410d7baead912e5658 not found: ID does not exist" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.365637 4823 scope.go:117] "RemoveContainer" containerID="f59365e42f4b2911a643e15c142b70cf04c844ac6f34d378dcc2fa6b002065f4" Mar 20 09:11:51 crc kubenswrapper[4823]: E0320 09:11:51.366169 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f59365e42f4b2911a643e15c142b70cf04c844ac6f34d378dcc2fa6b002065f4\": container with ID starting with f59365e42f4b2911a643e15c142b70cf04c844ac6f34d378dcc2fa6b002065f4 not found: ID does not exist" containerID="f59365e42f4b2911a643e15c142b70cf04c844ac6f34d378dcc2fa6b002065f4" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.366242 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f59365e42f4b2911a643e15c142b70cf04c844ac6f34d378dcc2fa6b002065f4"} err="failed to get container status \"f59365e42f4b2911a643e15c142b70cf04c844ac6f34d378dcc2fa6b002065f4\": rpc error: code = NotFound desc = could not find container \"f59365e42f4b2911a643e15c142b70cf04c844ac6f34d378dcc2fa6b002065f4\": container with ID starting with f59365e42f4b2911a643e15c142b70cf04c844ac6f34d378dcc2fa6b002065f4 not found: ID does not exist" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.366276 4823 scope.go:117] "RemoveContainer" containerID="eb05340d441f502026f1a83628d5230732cd88f064226103d7685d668408de36" Mar 20 09:11:51 crc kubenswrapper[4823]: E0320 09:11:51.366685 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb05340d441f502026f1a83628d5230732cd88f064226103d7685d668408de36\": container with ID starting with eb05340d441f502026f1a83628d5230732cd88f064226103d7685d668408de36 not found: ID does not exist" containerID="eb05340d441f502026f1a83628d5230732cd88f064226103d7685d668408de36" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.366719 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb05340d441f502026f1a83628d5230732cd88f064226103d7685d668408de36"} err="failed to get container status \"eb05340d441f502026f1a83628d5230732cd88f064226103d7685d668408de36\": rpc error: code = NotFound desc = could not find container \"eb05340d441f502026f1a83628d5230732cd88f064226103d7685d668408de36\": container with ID starting with eb05340d441f502026f1a83628d5230732cd88f064226103d7685d668408de36 not found: ID does not exist" Mar 20 09:11:51 crc kubenswrapper[4823]: I0320 09:11:51.968656 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6998581-faf4-4ab5-bc20-e1f0467f881d" (UID: "b6998581-faf4-4ab5-bc20-e1f0467f881d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:11:52 crc kubenswrapper[4823]: I0320 09:11:52.006387 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6998581-faf4-4ab5-bc20-e1f0467f881d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:52 crc kubenswrapper[4823]: I0320 09:11:52.219173 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hbn4z"] Mar 20 09:11:52 crc kubenswrapper[4823]: I0320 09:11:52.223111 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hbn4z"] Mar 20 09:11:52 crc kubenswrapper[4823]: I0320 09:11:52.695709 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-rm282" Mar 20 09:11:52 crc kubenswrapper[4823]: I0320 09:11:52.996315 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:52 crc kubenswrapper[4823]: I0320 09:11:52.996415 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:53 crc kubenswrapper[4823]: I0320 09:11:53.004121 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:53 crc kubenswrapper[4823]: I0320 09:11:53.318552 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-58d6f8d54f-znrsb" Mar 20 09:11:53 crc kubenswrapper[4823]: I0320 09:11:53.377040 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-lz4n9"] Mar 20 09:11:54 crc kubenswrapper[4823]: I0320 09:11:54.104145 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6998581-faf4-4ab5-bc20-e1f0467f881d" path="/var/lib/kubelet/pods/b6998581-faf4-4ab5-bc20-e1f0467f881d/volumes" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.129185 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566632-6mjng"] Mar 20 09:12:00 crc kubenswrapper[4823]: E0320 09:12:00.130071 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6998581-faf4-4ab5-bc20-e1f0467f881d" containerName="extract-utilities" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.130089 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6998581-faf4-4ab5-bc20-e1f0467f881d" containerName="extract-utilities" Mar 20 09:12:00 crc kubenswrapper[4823]: E0320 09:12:00.130113 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6998581-faf4-4ab5-bc20-e1f0467f881d" containerName="extract-content" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.130121 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6998581-faf4-4ab5-bc20-e1f0467f881d" containerName="extract-content" Mar 20 09:12:00 crc kubenswrapper[4823]: E0320 09:12:00.130135 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6998581-faf4-4ab5-bc20-e1f0467f881d" containerName="registry-server" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.130143 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6998581-faf4-4ab5-bc20-e1f0467f881d" containerName="registry-server" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.130272 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6998581-faf4-4ab5-bc20-e1f0467f881d" containerName="registry-server" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.130716 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566632-6mjng" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.133568 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.133900 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.135617 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.146378 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566632-6mjng"] Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.325579 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7twm2\" (UniqueName: \"kubernetes.io/projected/3e23c5c6-be76-47f5-baf3-8cfa530dea27-kube-api-access-7twm2\") pod \"auto-csr-approver-29566632-6mjng\" (UID: \"3e23c5c6-be76-47f5-baf3-8cfa530dea27\") " pod="openshift-infra/auto-csr-approver-29566632-6mjng" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.427932 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7twm2\" (UniqueName: \"kubernetes.io/projected/3e23c5c6-be76-47f5-baf3-8cfa530dea27-kube-api-access-7twm2\") pod \"auto-csr-approver-29566632-6mjng\" (UID: \"3e23c5c6-be76-47f5-baf3-8cfa530dea27\") " pod="openshift-infra/auto-csr-approver-29566632-6mjng" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.455280 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7twm2\" (UniqueName: \"kubernetes.io/projected/3e23c5c6-be76-47f5-baf3-8cfa530dea27-kube-api-access-7twm2\") pod \"auto-csr-approver-29566632-6mjng\" (UID: \"3e23c5c6-be76-47f5-baf3-8cfa530dea27\") " pod="openshift-infra/auto-csr-approver-29566632-6mjng" Mar 20 09:12:00 crc kubenswrapper[4823]: I0320 09:12:00.753969 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566632-6mjng" Mar 20 09:12:01 crc kubenswrapper[4823]: I0320 09:12:01.195729 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566632-6mjng"] Mar 20 09:12:01 crc kubenswrapper[4823]: I0320 09:12:01.363806 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566632-6mjng" event={"ID":"3e23c5c6-be76-47f5-baf3-8cfa530dea27","Type":"ContainerStarted","Data":"fbe41375119ae9a51fbcde2bbf6e839cea918ef475984a0c291968a14fb321de"} Mar 20 09:12:02 crc kubenswrapper[4823]: I0320 09:12:02.373592 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566632-6mjng" event={"ID":"3e23c5c6-be76-47f5-baf3-8cfa530dea27","Type":"ContainerStarted","Data":"bb63e63c35cd3a631abe26756abd61ed410a2fef3ce422b6c2b9c8dcea0a1b0b"} Mar 20 09:12:02 crc kubenswrapper[4823]: I0320 09:12:02.440109 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566632-6mjng" podStartSLOduration=1.542252833 podStartE2EDuration="2.44008853s" podCreationTimestamp="2026-03-20 09:12:00 +0000 UTC" firstStartedPulling="2026-03-20 09:12:01.206168887 +0000 UTC m=+887.892065291" lastFinishedPulling="2026-03-20 09:12:02.104004614 +0000 UTC m=+888.789900988" observedRunningTime="2026-03-20 09:12:02.437141039 +0000 UTC m=+889.123037413" watchObservedRunningTime="2026-03-20 09:12:02.44008853 +0000 UTC m=+889.125984904" Mar 20 09:12:03 crc kubenswrapper[4823]: I0320 09:12:03.234884 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-qfwd5" Mar 20 09:12:03 crc kubenswrapper[4823]: I0320 09:12:03.385934 4823 generic.go:334] "Generic (PLEG): container finished" podID="3e23c5c6-be76-47f5-baf3-8cfa530dea27" containerID="bb63e63c35cd3a631abe26756abd61ed410a2fef3ce422b6c2b9c8dcea0a1b0b" exitCode=0 Mar 20 09:12:03 crc kubenswrapper[4823]: I0320 09:12:03.385999 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566632-6mjng" event={"ID":"3e23c5c6-be76-47f5-baf3-8cfa530dea27","Type":"ContainerDied","Data":"bb63e63c35cd3a631abe26756abd61ed410a2fef3ce422b6c2b9c8dcea0a1b0b"} Mar 20 09:12:04 crc kubenswrapper[4823]: I0320 09:12:04.624316 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566632-6mjng" Mar 20 09:12:04 crc kubenswrapper[4823]: I0320 09:12:04.790612 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7twm2\" (UniqueName: \"kubernetes.io/projected/3e23c5c6-be76-47f5-baf3-8cfa530dea27-kube-api-access-7twm2\") pod \"3e23c5c6-be76-47f5-baf3-8cfa530dea27\" (UID: \"3e23c5c6-be76-47f5-baf3-8cfa530dea27\") " Mar 20 09:12:04 crc kubenswrapper[4823]: I0320 09:12:04.797377 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e23c5c6-be76-47f5-baf3-8cfa530dea27-kube-api-access-7twm2" (OuterVolumeSpecName: "kube-api-access-7twm2") pod "3e23c5c6-be76-47f5-baf3-8cfa530dea27" (UID: "3e23c5c6-be76-47f5-baf3-8cfa530dea27"). InnerVolumeSpecName "kube-api-access-7twm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:12:04 crc kubenswrapper[4823]: I0320 09:12:04.892484 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7twm2\" (UniqueName: \"kubernetes.io/projected/3e23c5c6-be76-47f5-baf3-8cfa530dea27-kube-api-access-7twm2\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:05 crc kubenswrapper[4823]: I0320 09:12:05.408931 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566632-6mjng" event={"ID":"3e23c5c6-be76-47f5-baf3-8cfa530dea27","Type":"ContainerDied","Data":"fbe41375119ae9a51fbcde2bbf6e839cea918ef475984a0c291968a14fb321de"} Mar 20 09:12:05 crc kubenswrapper[4823]: I0320 09:12:05.409022 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbe41375119ae9a51fbcde2bbf6e839cea918ef475984a0c291968a14fb321de" Mar 20 09:12:05 crc kubenswrapper[4823]: I0320 09:12:05.409108 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566632-6mjng" Mar 20 09:12:05 crc kubenswrapper[4823]: I0320 09:12:05.431960 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566626-z4qpc"] Mar 20 09:12:05 crc kubenswrapper[4823]: I0320 09:12:05.435488 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566626-z4qpc"] Mar 20 09:12:06 crc kubenswrapper[4823]: I0320 09:12:06.109382 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cfbe78d-6bcf-4fc2-b83e-e410c9af2522" path="/var/lib/kubelet/pods/0cfbe78d-6bcf-4fc2-b83e-e410c9af2522/volumes" Mar 20 09:12:14 crc kubenswrapper[4823]: I0320 09:12:14.823209 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc"] Mar 20 09:12:14 crc kubenswrapper[4823]: E0320 09:12:14.823961 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e23c5c6-be76-47f5-baf3-8cfa530dea27" containerName="oc" Mar 20 09:12:14 crc kubenswrapper[4823]: I0320 09:12:14.823976 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e23c5c6-be76-47f5-baf3-8cfa530dea27" containerName="oc" Mar 20 09:12:14 crc kubenswrapper[4823]: I0320 09:12:14.824141 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e23c5c6-be76-47f5-baf3-8cfa530dea27" containerName="oc" Mar 20 09:12:14 crc kubenswrapper[4823]: I0320 09:12:14.825149 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:14 crc kubenswrapper[4823]: I0320 09:12:14.826931 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 20 09:12:14 crc kubenswrapper[4823]: I0320 09:12:14.840598 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc"] Mar 20 09:12:14 crc kubenswrapper[4823]: I0320 09:12:14.924101 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:14 crc kubenswrapper[4823]: I0320 09:12:14.924176 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd577\" (UniqueName: \"kubernetes.io/projected/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-kube-api-access-wd577\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:14 crc kubenswrapper[4823]: I0320 09:12:14.924208 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:15 crc kubenswrapper[4823]: I0320 09:12:15.025470 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd577\" (UniqueName: \"kubernetes.io/projected/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-kube-api-access-wd577\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:15 crc kubenswrapper[4823]: I0320 09:12:15.025521 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:15 crc kubenswrapper[4823]: I0320 09:12:15.025600 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:15 crc kubenswrapper[4823]: I0320 09:12:15.026387 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:15 crc kubenswrapper[4823]: I0320 09:12:15.026398 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:15 crc kubenswrapper[4823]: I0320 09:12:15.042843 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd577\" (UniqueName: \"kubernetes.io/projected/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-kube-api-access-wd577\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:15 crc kubenswrapper[4823]: I0320 09:12:15.141927 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:15 crc kubenswrapper[4823]: I0320 09:12:15.668660 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc"] Mar 20 09:12:15 crc kubenswrapper[4823]: W0320 09:12:15.672679 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33cd5d1c_5e85_4f7b_a322_b78cf13812d5.slice/crio-7b94a5b751896c4e8dc8a7537cee96142fc15b54712c46b43e52a628ce600e68 WatchSource:0}: Error finding container 7b94a5b751896c4e8dc8a7537cee96142fc15b54712c46b43e52a628ce600e68: Status 404 returned error can't find the container with id 7b94a5b751896c4e8dc8a7537cee96142fc15b54712c46b43e52a628ce600e68 Mar 20 09:12:16 crc kubenswrapper[4823]: I0320 09:12:16.495635 4823 generic.go:334] "Generic (PLEG): container finished" podID="33cd5d1c-5e85-4f7b-a322-b78cf13812d5" containerID="f19671415100e4ff82ed990d887efa61e892e556403e4317b44491a1c2b0e5c3" exitCode=0 Mar 20 09:12:16 crc kubenswrapper[4823]: I0320 09:12:16.495697 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" event={"ID":"33cd5d1c-5e85-4f7b-a322-b78cf13812d5","Type":"ContainerDied","Data":"f19671415100e4ff82ed990d887efa61e892e556403e4317b44491a1c2b0e5c3"} Mar 20 09:12:16 crc kubenswrapper[4823]: I0320 09:12:16.495784 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" event={"ID":"33cd5d1c-5e85-4f7b-a322-b78cf13812d5","Type":"ContainerStarted","Data":"7b94a5b751896c4e8dc8a7537cee96142fc15b54712c46b43e52a628ce600e68"} Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.425025 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-lz4n9" podUID="e104c875-7420-45b3-b675-a63b92c454e4" containerName="console" containerID="cri-o://9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be" gracePeriod=15 Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.509439 4823 generic.go:334] "Generic (PLEG): container finished" podID="33cd5d1c-5e85-4f7b-a322-b78cf13812d5" containerID="21406c3f794340d615f8677b4c394b2730c874ae5bc628c88ecca971e73b5390" exitCode=0 Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.509486 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" event={"ID":"33cd5d1c-5e85-4f7b-a322-b78cf13812d5","Type":"ContainerDied","Data":"21406c3f794340d615f8677b4c394b2730c874ae5bc628c88ecca971e73b5390"} Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.781702 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-lz4n9_e104c875-7420-45b3-b675-a63b92c454e4/console/0.log" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.781790 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.888154 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-console-config\") pod \"e104c875-7420-45b3-b675-a63b92c454e4\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.888198 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-oauth-serving-cert\") pod \"e104c875-7420-45b3-b675-a63b92c454e4\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.888241 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-serving-cert\") pod \"e104c875-7420-45b3-b675-a63b92c454e4\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.888255 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46s55\" (UniqueName: \"kubernetes.io/projected/e104c875-7420-45b3-b675-a63b92c454e4-kube-api-access-46s55\") pod \"e104c875-7420-45b3-b675-a63b92c454e4\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.888313 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-oauth-config\") pod \"e104c875-7420-45b3-b675-a63b92c454e4\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.888341 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-service-ca\") pod \"e104c875-7420-45b3-b675-a63b92c454e4\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.888423 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-trusted-ca-bundle\") pod \"e104c875-7420-45b3-b675-a63b92c454e4\" (UID: \"e104c875-7420-45b3-b675-a63b92c454e4\") " Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.889004 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-service-ca" (OuterVolumeSpecName: "service-ca") pod "e104c875-7420-45b3-b675-a63b92c454e4" (UID: "e104c875-7420-45b3-b675-a63b92c454e4"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.889015 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-console-config" (OuterVolumeSpecName: "console-config") pod "e104c875-7420-45b3-b675-a63b92c454e4" (UID: "e104c875-7420-45b3-b675-a63b92c454e4"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.889109 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e104c875-7420-45b3-b675-a63b92c454e4" (UID: "e104c875-7420-45b3-b675-a63b92c454e4"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.889498 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e104c875-7420-45b3-b675-a63b92c454e4" (UID: "e104c875-7420-45b3-b675-a63b92c454e4"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.893565 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e104c875-7420-45b3-b675-a63b92c454e4" (UID: "e104c875-7420-45b3-b675-a63b92c454e4"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.893626 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e104c875-7420-45b3-b675-a63b92c454e4-kube-api-access-46s55" (OuterVolumeSpecName: "kube-api-access-46s55") pod "e104c875-7420-45b3-b675-a63b92c454e4" (UID: "e104c875-7420-45b3-b675-a63b92c454e4"). InnerVolumeSpecName "kube-api-access-46s55". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.893908 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e104c875-7420-45b3-b675-a63b92c454e4" (UID: "e104c875-7420-45b3-b675-a63b92c454e4"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.989619 4823 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.989661 4823 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-console-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.989672 4823 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.989683 4823 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.989692 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46s55\" (UniqueName: \"kubernetes.io/projected/e104c875-7420-45b3-b675-a63b92c454e4-kube-api-access-46s55\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.989705 4823 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e104c875-7420-45b3-b675-a63b92c454e4-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:18 crc kubenswrapper[4823]: I0320 09:12:18.989714 4823 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e104c875-7420-45b3-b675-a63b92c454e4-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.519081 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-lz4n9_e104c875-7420-45b3-b675-a63b92c454e4/console/0.log" Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.519144 4823 generic.go:334] "Generic (PLEG): container finished" podID="e104c875-7420-45b3-b675-a63b92c454e4" containerID="9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be" exitCode=2 Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.519232 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lz4n9" event={"ID":"e104c875-7420-45b3-b675-a63b92c454e4","Type":"ContainerDied","Data":"9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be"} Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.519283 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-lz4n9" event={"ID":"e104c875-7420-45b3-b675-a63b92c454e4","Type":"ContainerDied","Data":"006f746577c69e05f0160fea69f9a67239dcac7a2446d93eac85c8f7fcd34af2"} Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.519245 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-lz4n9" Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.519301 4823 scope.go:117] "RemoveContainer" containerID="9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be" Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.524188 4823 generic.go:334] "Generic (PLEG): container finished" podID="33cd5d1c-5e85-4f7b-a322-b78cf13812d5" containerID="7444b1834d9f827274403ce08979071aba74625c8664cefc71423b98583c19dc" exitCode=0 Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.524246 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" event={"ID":"33cd5d1c-5e85-4f7b-a322-b78cf13812d5","Type":"ContainerDied","Data":"7444b1834d9f827274403ce08979071aba74625c8664cefc71423b98583c19dc"} Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.541182 4823 scope.go:117] "RemoveContainer" containerID="9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be" Mar 20 09:12:19 crc kubenswrapper[4823]: E0320 09:12:19.544121 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be\": container with ID starting with 9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be not found: ID does not exist" containerID="9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be" Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.544155 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be"} err="failed to get container status \"9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be\": rpc error: code = NotFound desc = could not find container \"9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be\": container with ID starting with 9d8c7fea4331f94f6d7e3b72451fcc511c81fd59815ac8901335cbc452de07be not found: ID does not exist" Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.561593 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-lz4n9"] Mar 20 09:12:19 crc kubenswrapper[4823]: I0320 09:12:19.567341 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-lz4n9"] Mar 20 09:12:20 crc kubenswrapper[4823]: I0320 09:12:20.107002 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e104c875-7420-45b3-b675-a63b92c454e4" path="/var/lib/kubelet/pods/e104c875-7420-45b3-b675-a63b92c454e4/volumes" Mar 20 09:12:20 crc kubenswrapper[4823]: I0320 09:12:20.808159 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:20 crc kubenswrapper[4823]: I0320 09:12:20.914248 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd577\" (UniqueName: \"kubernetes.io/projected/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-kube-api-access-wd577\") pod \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " Mar 20 09:12:20 crc kubenswrapper[4823]: I0320 09:12:20.914317 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-bundle\") pod \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " Mar 20 09:12:20 crc kubenswrapper[4823]: I0320 09:12:20.914365 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-util\") pod \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\" (UID: \"33cd5d1c-5e85-4f7b-a322-b78cf13812d5\") " Mar 20 09:12:20 crc kubenswrapper[4823]: I0320 09:12:20.915655 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-bundle" (OuterVolumeSpecName: "bundle") pod "33cd5d1c-5e85-4f7b-a322-b78cf13812d5" (UID: "33cd5d1c-5e85-4f7b-a322-b78cf13812d5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:12:20 crc kubenswrapper[4823]: I0320 09:12:20.920343 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-kube-api-access-wd577" (OuterVolumeSpecName: "kube-api-access-wd577") pod "33cd5d1c-5e85-4f7b-a322-b78cf13812d5" (UID: "33cd5d1c-5e85-4f7b-a322-b78cf13812d5"). InnerVolumeSpecName "kube-api-access-wd577". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:12:20 crc kubenswrapper[4823]: I0320 09:12:20.944660 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-util" (OuterVolumeSpecName: "util") pod "33cd5d1c-5e85-4f7b-a322-b78cf13812d5" (UID: "33cd5d1c-5e85-4f7b-a322-b78cf13812d5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:12:21 crc kubenswrapper[4823]: I0320 09:12:21.016383 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd577\" (UniqueName: \"kubernetes.io/projected/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-kube-api-access-wd577\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:21 crc kubenswrapper[4823]: I0320 09:12:21.016427 4823 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:21 crc kubenswrapper[4823]: I0320 09:12:21.016439 4823 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33cd5d1c-5e85-4f7b-a322-b78cf13812d5-util\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:21 crc kubenswrapper[4823]: I0320 09:12:21.540520 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" event={"ID":"33cd5d1c-5e85-4f7b-a322-b78cf13812d5","Type":"ContainerDied","Data":"7b94a5b751896c4e8dc8a7537cee96142fc15b54712c46b43e52a628ce600e68"} Mar 20 09:12:21 crc kubenswrapper[4823]: I0320 09:12:21.540560 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b94a5b751896c4e8dc8a7537cee96142fc15b54712c46b43e52a628ce600e68" Mar 20 09:12:21 crc kubenswrapper[4823]: I0320 09:12:21.540628 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.069304 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2"] Mar 20 09:12:30 crc kubenswrapper[4823]: E0320 09:12:30.070059 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e104c875-7420-45b3-b675-a63b92c454e4" containerName="console" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.070076 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="e104c875-7420-45b3-b675-a63b92c454e4" containerName="console" Mar 20 09:12:30 crc kubenswrapper[4823]: E0320 09:12:30.070091 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33cd5d1c-5e85-4f7b-a322-b78cf13812d5" containerName="pull" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.070098 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="33cd5d1c-5e85-4f7b-a322-b78cf13812d5" containerName="pull" Mar 20 09:12:30 crc kubenswrapper[4823]: E0320 09:12:30.070115 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33cd5d1c-5e85-4f7b-a322-b78cf13812d5" containerName="util" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.070124 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="33cd5d1c-5e85-4f7b-a322-b78cf13812d5" containerName="util" Mar 20 09:12:30 crc kubenswrapper[4823]: E0320 09:12:30.070136 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33cd5d1c-5e85-4f7b-a322-b78cf13812d5" containerName="extract" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.070143 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="33cd5d1c-5e85-4f7b-a322-b78cf13812d5" containerName="extract" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.070283 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="33cd5d1c-5e85-4f7b-a322-b78cf13812d5" containerName="extract" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.070294 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="e104c875-7420-45b3-b675-a63b92c454e4" containerName="console" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.070765 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.077048 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.077121 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.077069 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-r6775" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.077260 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.077499 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.119983 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2"] Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.230462 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz695\" (UniqueName: \"kubernetes.io/projected/d4d5d92a-98ad-40ff-9b4e-4568375f78b4-kube-api-access-gz695\") pod \"metallb-operator-controller-manager-855dc6dbf9-tw2n2\" (UID: \"d4d5d92a-98ad-40ff-9b4e-4568375f78b4\") " pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.230510 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d4d5d92a-98ad-40ff-9b4e-4568375f78b4-apiservice-cert\") pod \"metallb-operator-controller-manager-855dc6dbf9-tw2n2\" (UID: \"d4d5d92a-98ad-40ff-9b4e-4568375f78b4\") " pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.230659 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d4d5d92a-98ad-40ff-9b4e-4568375f78b4-webhook-cert\") pod \"metallb-operator-controller-manager-855dc6dbf9-tw2n2\" (UID: \"d4d5d92a-98ad-40ff-9b4e-4568375f78b4\") " pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.332344 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d4d5d92a-98ad-40ff-9b4e-4568375f78b4-webhook-cert\") pod \"metallb-operator-controller-manager-855dc6dbf9-tw2n2\" (UID: \"d4d5d92a-98ad-40ff-9b4e-4568375f78b4\") " pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.332672 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz695\" (UniqueName: \"kubernetes.io/projected/d4d5d92a-98ad-40ff-9b4e-4568375f78b4-kube-api-access-gz695\") pod \"metallb-operator-controller-manager-855dc6dbf9-tw2n2\" (UID: \"d4d5d92a-98ad-40ff-9b4e-4568375f78b4\") " pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.332694 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d4d5d92a-98ad-40ff-9b4e-4568375f78b4-apiservice-cert\") pod \"metallb-operator-controller-manager-855dc6dbf9-tw2n2\" (UID: \"d4d5d92a-98ad-40ff-9b4e-4568375f78b4\") " pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.350934 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d4d5d92a-98ad-40ff-9b4e-4568375f78b4-webhook-cert\") pod \"metallb-operator-controller-manager-855dc6dbf9-tw2n2\" (UID: \"d4d5d92a-98ad-40ff-9b4e-4568375f78b4\") " pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.351074 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d4d5d92a-98ad-40ff-9b4e-4568375f78b4-apiservice-cert\") pod \"metallb-operator-controller-manager-855dc6dbf9-tw2n2\" (UID: \"d4d5d92a-98ad-40ff-9b4e-4568375f78b4\") " pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.354811 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz695\" (UniqueName: \"kubernetes.io/projected/d4d5d92a-98ad-40ff-9b4e-4568375f78b4-kube-api-access-gz695\") pod \"metallb-operator-controller-manager-855dc6dbf9-tw2n2\" (UID: \"d4d5d92a-98ad-40ff-9b4e-4568375f78b4\") " pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.390754 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.400903 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4"] Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.401798 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.403970 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-vtmjg" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.404119 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.404645 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.413567 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4"] Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.534931 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl6d8\" (UniqueName: \"kubernetes.io/projected/f6438eac-8065-43a7-9c1a-eabc0a7ac277-kube-api-access-vl6d8\") pod \"metallb-operator-webhook-server-5d4db96fb4-qjtz4\" (UID: \"f6438eac-8065-43a7-9c1a-eabc0a7ac277\") " pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.535334 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6438eac-8065-43a7-9c1a-eabc0a7ac277-apiservice-cert\") pod \"metallb-operator-webhook-server-5d4db96fb4-qjtz4\" (UID: \"f6438eac-8065-43a7-9c1a-eabc0a7ac277\") " pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.535396 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6438eac-8065-43a7-9c1a-eabc0a7ac277-webhook-cert\") pod \"metallb-operator-webhook-server-5d4db96fb4-qjtz4\" (UID: \"f6438eac-8065-43a7-9c1a-eabc0a7ac277\") " pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.628490 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2"] Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.637120 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6438eac-8065-43a7-9c1a-eabc0a7ac277-webhook-cert\") pod \"metallb-operator-webhook-server-5d4db96fb4-qjtz4\" (UID: \"f6438eac-8065-43a7-9c1a-eabc0a7ac277\") " pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.637212 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl6d8\" (UniqueName: \"kubernetes.io/projected/f6438eac-8065-43a7-9c1a-eabc0a7ac277-kube-api-access-vl6d8\") pod \"metallb-operator-webhook-server-5d4db96fb4-qjtz4\" (UID: \"f6438eac-8065-43a7-9c1a-eabc0a7ac277\") " pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.637259 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6438eac-8065-43a7-9c1a-eabc0a7ac277-apiservice-cert\") pod \"metallb-operator-webhook-server-5d4db96fb4-qjtz4\" (UID: \"f6438eac-8065-43a7-9c1a-eabc0a7ac277\") " pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.643679 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6438eac-8065-43a7-9c1a-eabc0a7ac277-webhook-cert\") pod \"metallb-operator-webhook-server-5d4db96fb4-qjtz4\" (UID: \"f6438eac-8065-43a7-9c1a-eabc0a7ac277\") " pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.653418 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6438eac-8065-43a7-9c1a-eabc0a7ac277-apiservice-cert\") pod \"metallb-operator-webhook-server-5d4db96fb4-qjtz4\" (UID: \"f6438eac-8065-43a7-9c1a-eabc0a7ac277\") " pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.668449 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl6d8\" (UniqueName: \"kubernetes.io/projected/f6438eac-8065-43a7-9c1a-eabc0a7ac277-kube-api-access-vl6d8\") pod \"metallb-operator-webhook-server-5d4db96fb4-qjtz4\" (UID: \"f6438eac-8065-43a7-9c1a-eabc0a7ac277\") " pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.755316 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:30 crc kubenswrapper[4823]: I0320 09:12:30.969813 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4"] Mar 20 09:12:30 crc kubenswrapper[4823]: W0320 09:12:30.976107 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6438eac_8065_43a7_9c1a_eabc0a7ac277.slice/crio-31c830a97087b4aca58663c0c281ff11a6b69eb39206f37a0ab1a4463cb5c4e1 WatchSource:0}: Error finding container 31c830a97087b4aca58663c0c281ff11a6b69eb39206f37a0ab1a4463cb5c4e1: Status 404 returned error can't find the container with id 31c830a97087b4aca58663c0c281ff11a6b69eb39206f37a0ab1a4463cb5c4e1 Mar 20 09:12:31 crc kubenswrapper[4823]: I0320 09:12:31.579201 4823 scope.go:117] "RemoveContainer" containerID="02b48ffff6a5ae7706309c9d6a0252802210cdf945cf5637687d3ac055aa10df" Mar 20 09:12:31 crc kubenswrapper[4823]: I0320 09:12:31.598207 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" event={"ID":"d4d5d92a-98ad-40ff-9b4e-4568375f78b4","Type":"ContainerStarted","Data":"32a3ddbfd46f50782329739644f0980cf229912b2e141829733436652c85a085"} Mar 20 09:12:31 crc kubenswrapper[4823]: I0320 09:12:31.599379 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" event={"ID":"f6438eac-8065-43a7-9c1a-eabc0a7ac277","Type":"ContainerStarted","Data":"31c830a97087b4aca58663c0c281ff11a6b69eb39206f37a0ab1a4463cb5c4e1"} Mar 20 09:12:36 crc kubenswrapper[4823]: I0320 09:12:36.636665 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" event={"ID":"d4d5d92a-98ad-40ff-9b4e-4568375f78b4","Type":"ContainerStarted","Data":"714fde2e59c717b958806f1f592123edbd153a5a6990e4c3950df090a6f272b5"} Mar 20 09:12:36 crc kubenswrapper[4823]: I0320 09:12:36.637253 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:12:36 crc kubenswrapper[4823]: I0320 09:12:36.638002 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" event={"ID":"f6438eac-8065-43a7-9c1a-eabc0a7ac277","Type":"ContainerStarted","Data":"1731b5b43ef87816cab34529920027103cabd8fce9d852e902ecc470a8b75748"} Mar 20 09:12:36 crc kubenswrapper[4823]: I0320 09:12:36.638162 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:12:36 crc kubenswrapper[4823]: I0320 09:12:36.656939 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" podStartSLOduration=1.2002587120000001 podStartE2EDuration="6.656925583s" podCreationTimestamp="2026-03-20 09:12:30 +0000 UTC" firstStartedPulling="2026-03-20 09:12:30.66850692 +0000 UTC m=+917.354403294" lastFinishedPulling="2026-03-20 09:12:36.125173801 +0000 UTC m=+922.811070165" observedRunningTime="2026-03-20 09:12:36.654318122 +0000 UTC m=+923.340214496" watchObservedRunningTime="2026-03-20 09:12:36.656925583 +0000 UTC m=+923.342821957" Mar 20 09:12:36 crc kubenswrapper[4823]: I0320 09:12:36.679228 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" podStartSLOduration=1.5319432960000001 podStartE2EDuration="6.679195517s" podCreationTimestamp="2026-03-20 09:12:30 +0000 UTC" firstStartedPulling="2026-03-20 09:12:30.981873244 +0000 UTC m=+917.667769628" lastFinishedPulling="2026-03-20 09:12:36.129125475 +0000 UTC m=+922.815021849" observedRunningTime="2026-03-20 09:12:36.674450989 +0000 UTC m=+923.360347363" watchObservedRunningTime="2026-03-20 09:12:36.679195517 +0000 UTC m=+923.365091891" Mar 20 09:12:50 crc kubenswrapper[4823]: I0320 09:12:50.759266 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5d4db96fb4-qjtz4" Mar 20 09:13:09 crc kubenswrapper[4823]: I0320 09:13:09.430330 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:13:09 crc kubenswrapper[4823]: I0320 09:13:09.430896 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:13:10 crc kubenswrapper[4823]: I0320 09:13:10.393582 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-855dc6dbf9-tw2n2" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.111915 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-zgtrw"] Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.125200 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.127193 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.127225 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.127564 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-cm278" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.128400 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd"] Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.129214 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.132079 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.142259 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd"] Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.197640 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-cmc9m"] Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.198506 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.202243 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.202359 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.202468 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-7bwwr" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.202593 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.225194 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-jh928"] Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.226368 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.229632 4823 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.235064 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-jh928"] Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.289337 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-reloader\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.289381 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-frr-sockets\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.289412 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-frr-startup\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.289430 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtwkh\" (UniqueName: \"kubernetes.io/projected/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-kube-api-access-jtwkh\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.289465 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-frr-conf\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.289490 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-metrics\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.289994 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrqzb\" (UniqueName: \"kubernetes.io/projected/ef072780-7e18-4446-b701-501534f52d8d-kube-api-access-wrqzb\") pod \"frr-k8s-webhook-server-bcc4b6f68-trvvd\" (UID: \"ef072780-7e18-4446-b701-501534f52d8d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.290094 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ef072780-7e18-4446-b701-501534f52d8d-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-trvvd\" (UID: \"ef072780-7e18-4446-b701-501534f52d8d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.290270 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-metrics-certs\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.391897 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-memberlist\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.391954 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-reloader\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.391979 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ef19c0e0-efc5-4902-beb9-99b865a664c1-metallb-excludel2\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.391995 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1de94133-3c13-40f1-980f-da1342011a04-cert\") pod \"controller-7bb4cc7c98-jh928\" (UID: \"1de94133-3c13-40f1-980f-da1342011a04\") " pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392037 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-frr-sockets\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392062 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-frr-startup\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392076 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtwkh\" (UniqueName: \"kubernetes.io/projected/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-kube-api-access-jtwkh\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392100 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g4hj\" (UniqueName: \"kubernetes.io/projected/ef19c0e0-efc5-4902-beb9-99b865a664c1-kube-api-access-8g4hj\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392123 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-metrics-certs\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392137 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8ss7\" (UniqueName: \"kubernetes.io/projected/1de94133-3c13-40f1-980f-da1342011a04-kube-api-access-l8ss7\") pod \"controller-7bb4cc7c98-jh928\" (UID: \"1de94133-3c13-40f1-980f-da1342011a04\") " pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392155 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-frr-conf\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392172 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-metrics\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392201 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrqzb\" (UniqueName: \"kubernetes.io/projected/ef072780-7e18-4446-b701-501534f52d8d-kube-api-access-wrqzb\") pod \"frr-k8s-webhook-server-bcc4b6f68-trvvd\" (UID: \"ef072780-7e18-4446-b701-501534f52d8d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392220 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1de94133-3c13-40f1-980f-da1342011a04-metrics-certs\") pod \"controller-7bb4cc7c98-jh928\" (UID: \"1de94133-3c13-40f1-980f-da1342011a04\") " pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392242 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ef072780-7e18-4446-b701-501534f52d8d-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-trvvd\" (UID: \"ef072780-7e18-4446-b701-501534f52d8d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392278 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-metrics-certs\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.392872 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-reloader\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.393043 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-metrics\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.393117 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-frr-sockets\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.393509 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-frr-conf\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.394008 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-frr-startup\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.399621 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ef072780-7e18-4446-b701-501534f52d8d-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-trvvd\" (UID: \"ef072780-7e18-4446-b701-501534f52d8d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.401232 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-metrics-certs\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.414144 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrqzb\" (UniqueName: \"kubernetes.io/projected/ef072780-7e18-4446-b701-501534f52d8d-kube-api-access-wrqzb\") pod \"frr-k8s-webhook-server-bcc4b6f68-trvvd\" (UID: \"ef072780-7e18-4446-b701-501534f52d8d\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.417318 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtwkh\" (UniqueName: \"kubernetes.io/projected/00eeb55c-4fd8-4aa3-ac81-b687c2074a5e-kube-api-access-jtwkh\") pod \"frr-k8s-zgtrw\" (UID: \"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e\") " pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.457494 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.469987 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.493606 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-memberlist\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.493656 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ef19c0e0-efc5-4902-beb9-99b865a664c1-metallb-excludel2\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.493678 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1de94133-3c13-40f1-980f-da1342011a04-cert\") pod \"controller-7bb4cc7c98-jh928\" (UID: \"1de94133-3c13-40f1-980f-da1342011a04\") " pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.493709 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g4hj\" (UniqueName: \"kubernetes.io/projected/ef19c0e0-efc5-4902-beb9-99b865a664c1-kube-api-access-8g4hj\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.493735 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-metrics-certs\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.493771 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8ss7\" (UniqueName: \"kubernetes.io/projected/1de94133-3c13-40f1-980f-da1342011a04-kube-api-access-l8ss7\") pod \"controller-7bb4cc7c98-jh928\" (UID: \"1de94133-3c13-40f1-980f-da1342011a04\") " pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.493808 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1de94133-3c13-40f1-980f-da1342011a04-metrics-certs\") pod \"controller-7bb4cc7c98-jh928\" (UID: \"1de94133-3c13-40f1-980f-da1342011a04\") " pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:11 crc kubenswrapper[4823]: E0320 09:13:11.494672 4823 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 20 09:13:11 crc kubenswrapper[4823]: E0320 09:13:11.494889 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-memberlist podName:ef19c0e0-efc5-4902-beb9-99b865a664c1 nodeName:}" failed. No retries permitted until 2026-03-20 09:13:11.994858047 +0000 UTC m=+958.680754461 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-memberlist") pod "speaker-cmc9m" (UID: "ef19c0e0-efc5-4902-beb9-99b865a664c1") : secret "metallb-memberlist" not found Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.494700 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ef19c0e0-efc5-4902-beb9-99b865a664c1-metallb-excludel2\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.499427 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1de94133-3c13-40f1-980f-da1342011a04-cert\") pod \"controller-7bb4cc7c98-jh928\" (UID: \"1de94133-3c13-40f1-980f-da1342011a04\") " pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.500330 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1de94133-3c13-40f1-980f-da1342011a04-metrics-certs\") pod \"controller-7bb4cc7c98-jh928\" (UID: \"1de94133-3c13-40f1-980f-da1342011a04\") " pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.501004 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-metrics-certs\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.510280 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8ss7\" (UniqueName: \"kubernetes.io/projected/1de94133-3c13-40f1-980f-da1342011a04-kube-api-access-l8ss7\") pod \"controller-7bb4cc7c98-jh928\" (UID: \"1de94133-3c13-40f1-980f-da1342011a04\") " pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.513608 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g4hj\" (UniqueName: \"kubernetes.io/projected/ef19c0e0-efc5-4902-beb9-99b865a664c1-kube-api-access-8g4hj\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.544193 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.860180 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zgtrw" event={"ID":"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e","Type":"ContainerStarted","Data":"0911edfa914a6835190e24ade1b1fd122fc36d3ed7f56a2b424573fcb14ce28e"} Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.872972 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd"] Mar 20 09:13:11 crc kubenswrapper[4823]: W0320 09:13:11.879254 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef072780_7e18_4446_b701_501534f52d8d.slice/crio-ed20e6455497d224f52a9804fb60a061bac51fc6f6b4ad6041deafa20c329371 WatchSource:0}: Error finding container ed20e6455497d224f52a9804fb60a061bac51fc6f6b4ad6041deafa20c329371: Status 404 returned error can't find the container with id ed20e6455497d224f52a9804fb60a061bac51fc6f6b4ad6041deafa20c329371 Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.954662 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-jh928"] Mar 20 09:13:11 crc kubenswrapper[4823]: W0320 09:13:11.960420 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1de94133_3c13_40f1_980f_da1342011a04.slice/crio-faac83d63de7bba064e22bc8e008d807035d800c874ad908d1a924e9b2cc7e7d WatchSource:0}: Error finding container faac83d63de7bba064e22bc8e008d807035d800c874ad908d1a924e9b2cc7e7d: Status 404 returned error can't find the container with id faac83d63de7bba064e22bc8e008d807035d800c874ad908d1a924e9b2cc7e7d Mar 20 09:13:11 crc kubenswrapper[4823]: I0320 09:13:11.998657 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-memberlist\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:11 crc kubenswrapper[4823]: E0320 09:13:11.998857 4823 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 20 09:13:11 crc kubenswrapper[4823]: E0320 09:13:11.998936 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-memberlist podName:ef19c0e0-efc5-4902-beb9-99b865a664c1 nodeName:}" failed. No retries permitted until 2026-03-20 09:13:12.998917187 +0000 UTC m=+959.684813561 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-memberlist") pod "speaker-cmc9m" (UID: "ef19c0e0-efc5-4902-beb9-99b865a664c1") : secret "metallb-memberlist" not found Mar 20 09:13:12 crc kubenswrapper[4823]: I0320 09:13:12.869666 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-jh928" event={"ID":"1de94133-3c13-40f1-980f-da1342011a04","Type":"ContainerStarted","Data":"abb3e583904a2b2961c70f33f3139cbca1626ba915b6ac5a106273198b55ac3b"} Mar 20 09:13:12 crc kubenswrapper[4823]: I0320 09:13:12.870015 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-jh928" event={"ID":"1de94133-3c13-40f1-980f-da1342011a04","Type":"ContainerStarted","Data":"d08890ac93b09fce3b4bfe7525a0283fb728f7a3bac8cc434800ed1fbf28f3fd"} Mar 20 09:13:12 crc kubenswrapper[4823]: I0320 09:13:12.870043 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:12 crc kubenswrapper[4823]: I0320 09:13:12.870057 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-jh928" event={"ID":"1de94133-3c13-40f1-980f-da1342011a04","Type":"ContainerStarted","Data":"faac83d63de7bba064e22bc8e008d807035d800c874ad908d1a924e9b2cc7e7d"} Mar 20 09:13:12 crc kubenswrapper[4823]: I0320 09:13:12.870974 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" event={"ID":"ef072780-7e18-4446-b701-501534f52d8d","Type":"ContainerStarted","Data":"ed20e6455497d224f52a9804fb60a061bac51fc6f6b4ad6041deafa20c329371"} Mar 20 09:13:12 crc kubenswrapper[4823]: I0320 09:13:12.895827 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-jh928" podStartSLOduration=1.895805825 podStartE2EDuration="1.895805825s" podCreationTimestamp="2026-03-20 09:13:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:13:12.891697767 +0000 UTC m=+959.577594161" watchObservedRunningTime="2026-03-20 09:13:12.895805825 +0000 UTC m=+959.581702199" Mar 20 09:13:13 crc kubenswrapper[4823]: I0320 09:13:13.010708 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-memberlist\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:13 crc kubenswrapper[4823]: I0320 09:13:13.026413 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ef19c0e0-efc5-4902-beb9-99b865a664c1-memberlist\") pod \"speaker-cmc9m\" (UID: \"ef19c0e0-efc5-4902-beb9-99b865a664c1\") " pod="metallb-system/speaker-cmc9m" Mar 20 09:13:13 crc kubenswrapper[4823]: I0320 09:13:13.311067 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cmc9m" Mar 20 09:13:13 crc kubenswrapper[4823]: I0320 09:13:13.877843 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cmc9m" event={"ID":"ef19c0e0-efc5-4902-beb9-99b865a664c1","Type":"ContainerStarted","Data":"e463a5cff288c02ccafb760930b7502402241b48d5d40fd346d49f017b2c4039"} Mar 20 09:13:13 crc kubenswrapper[4823]: I0320 09:13:13.878179 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cmc9m" event={"ID":"ef19c0e0-efc5-4902-beb9-99b865a664c1","Type":"ContainerStarted","Data":"e78ba6987115fd9b4bc46d94c5a12ef3d7136f1c56babef203f801b0f0804f5b"} Mar 20 09:13:13 crc kubenswrapper[4823]: I0320 09:13:13.878191 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cmc9m" event={"ID":"ef19c0e0-efc5-4902-beb9-99b865a664c1","Type":"ContainerStarted","Data":"0e0aa6c0f982202fcc4ce30a6e2648df56c49a8732ce164c5f295cd98e0c4e74"} Mar 20 09:13:13 crc kubenswrapper[4823]: I0320 09:13:13.878351 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-cmc9m" Mar 20 09:13:13 crc kubenswrapper[4823]: I0320 09:13:13.903310 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-cmc9m" podStartSLOduration=2.903293456 podStartE2EDuration="2.903293456s" podCreationTimestamp="2026-03-20 09:13:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:13:13.899512237 +0000 UTC m=+960.585408611" watchObservedRunningTime="2026-03-20 09:13:13.903293456 +0000 UTC m=+960.589189830" Mar 20 09:13:18 crc kubenswrapper[4823]: I0320 09:13:18.909087 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" event={"ID":"ef072780-7e18-4446-b701-501534f52d8d","Type":"ContainerStarted","Data":"315ecc23e6ae1c8ff26de1007e3892baf4565b31755350393ba4f5bbdc83eaf1"} Mar 20 09:13:18 crc kubenswrapper[4823]: I0320 09:13:18.909676 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" Mar 20 09:13:18 crc kubenswrapper[4823]: I0320 09:13:18.911874 4823 generic.go:334] "Generic (PLEG): container finished" podID="00eeb55c-4fd8-4aa3-ac81-b687c2074a5e" containerID="2c763b4d18e4466b71a5448ce5dfdb9f469384cc0672c86977d418b46dedf3bb" exitCode=0 Mar 20 09:13:18 crc kubenswrapper[4823]: I0320 09:13:18.911926 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zgtrw" event={"ID":"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e","Type":"ContainerDied","Data":"2c763b4d18e4466b71a5448ce5dfdb9f469384cc0672c86977d418b46dedf3bb"} Mar 20 09:13:18 crc kubenswrapper[4823]: I0320 09:13:18.945348 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" podStartSLOduration=1.267528058 podStartE2EDuration="7.94532418s" podCreationTimestamp="2026-03-20 09:13:11 +0000 UTC" firstStartedPulling="2026-03-20 09:13:11.882184834 +0000 UTC m=+958.568081208" lastFinishedPulling="2026-03-20 09:13:18.559980916 +0000 UTC m=+965.245877330" observedRunningTime="2026-03-20 09:13:18.940954665 +0000 UTC m=+965.626851039" watchObservedRunningTime="2026-03-20 09:13:18.94532418 +0000 UTC m=+965.631220564" Mar 20 09:13:19 crc kubenswrapper[4823]: I0320 09:13:19.920175 4823 generic.go:334] "Generic (PLEG): container finished" podID="00eeb55c-4fd8-4aa3-ac81-b687c2074a5e" containerID="be06bf2bb37e162fd0a7e82f9b8149dfeb9fa7f700f349916674f115cc769661" exitCode=0 Mar 20 09:13:19 crc kubenswrapper[4823]: I0320 09:13:19.921561 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zgtrw" event={"ID":"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e","Type":"ContainerDied","Data":"be06bf2bb37e162fd0a7e82f9b8149dfeb9fa7f700f349916674f115cc769661"} Mar 20 09:13:20 crc kubenswrapper[4823]: I0320 09:13:20.933673 4823 generic.go:334] "Generic (PLEG): container finished" podID="00eeb55c-4fd8-4aa3-ac81-b687c2074a5e" containerID="08399f6f419476412c30a770eee0f0f2df4c49e611cbac58c1ec6adbf8a586bb" exitCode=0 Mar 20 09:13:20 crc kubenswrapper[4823]: I0320 09:13:20.933768 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zgtrw" event={"ID":"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e","Type":"ContainerDied","Data":"08399f6f419476412c30a770eee0f0f2df4c49e611cbac58c1ec6adbf8a586bb"} Mar 20 09:13:21 crc kubenswrapper[4823]: I0320 09:13:21.944264 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zgtrw" event={"ID":"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e","Type":"ContainerStarted","Data":"d6256362a30dffc47db176d7dfa70a8095d77333a25e399b9faf22db02507aa5"} Mar 20 09:13:21 crc kubenswrapper[4823]: I0320 09:13:21.944586 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:21 crc kubenswrapper[4823]: I0320 09:13:21.944597 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zgtrw" event={"ID":"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e","Type":"ContainerStarted","Data":"d9ec598b7f4ec34f418bb78b2bdaf927a1f9a9cd2efc52adc75d2514fa0dbee5"} Mar 20 09:13:21 crc kubenswrapper[4823]: I0320 09:13:21.944607 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zgtrw" event={"ID":"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e","Type":"ContainerStarted","Data":"8c0771bd1e58e4dd4e4c8572d89c47239100437b2610253d8e4ad719585c7b6f"} Mar 20 09:13:21 crc kubenswrapper[4823]: I0320 09:13:21.944617 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zgtrw" event={"ID":"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e","Type":"ContainerStarted","Data":"03789ce35f7d4023cf1388bdb8635c3f6831a34f5f4efa09bab0a355cf860ca8"} Mar 20 09:13:21 crc kubenswrapper[4823]: I0320 09:13:21.944625 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zgtrw" event={"ID":"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e","Type":"ContainerStarted","Data":"2d74be21830ddbc6523fc7778ea61bc61739f5d39be7a08ede0531f9553bf509"} Mar 20 09:13:21 crc kubenswrapper[4823]: I0320 09:13:21.944636 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zgtrw" event={"ID":"00eeb55c-4fd8-4aa3-ac81-b687c2074a5e","Type":"ContainerStarted","Data":"f9fb0c456b7bf3b70b6536cec5017cf4680e1a70af34c97cdc5ecb24391d2eae"} Mar 20 09:13:21 crc kubenswrapper[4823]: I0320 09:13:21.967055 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-zgtrw" podStartSLOduration=3.999447277 podStartE2EDuration="10.967036759s" podCreationTimestamp="2026-03-20 09:13:11 +0000 UTC" firstStartedPulling="2026-03-20 09:13:11.623798491 +0000 UTC m=+958.309694865" lastFinishedPulling="2026-03-20 09:13:18.591387973 +0000 UTC m=+965.277284347" observedRunningTime="2026-03-20 09:13:21.963488298 +0000 UTC m=+968.649384672" watchObservedRunningTime="2026-03-20 09:13:21.967036759 +0000 UTC m=+968.652933133" Mar 20 09:13:23 crc kubenswrapper[4823]: I0320 09:13:23.315458 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-cmc9m" Mar 20 09:13:25 crc kubenswrapper[4823]: I0320 09:13:25.792063 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-lgzqc"] Mar 20 09:13:25 crc kubenswrapper[4823]: I0320 09:13:25.793460 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lgzqc" Mar 20 09:13:25 crc kubenswrapper[4823]: I0320 09:13:25.797353 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 20 09:13:25 crc kubenswrapper[4823]: I0320 09:13:25.807388 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 20 09:13:25 crc kubenswrapper[4823]: I0320 09:13:25.809276 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-z9mcw" Mar 20 09:13:25 crc kubenswrapper[4823]: I0320 09:13:25.818801 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lgzqc"] Mar 20 09:13:25 crc kubenswrapper[4823]: I0320 09:13:25.906516 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk9wh\" (UniqueName: \"kubernetes.io/projected/75433774-57f1-47f6-9472-d36aa486ea2e-kube-api-access-hk9wh\") pod \"openstack-operator-index-lgzqc\" (UID: \"75433774-57f1-47f6-9472-d36aa486ea2e\") " pod="openstack-operators/openstack-operator-index-lgzqc" Mar 20 09:13:26 crc kubenswrapper[4823]: I0320 09:13:26.007271 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk9wh\" (UniqueName: \"kubernetes.io/projected/75433774-57f1-47f6-9472-d36aa486ea2e-kube-api-access-hk9wh\") pod \"openstack-operator-index-lgzqc\" (UID: \"75433774-57f1-47f6-9472-d36aa486ea2e\") " pod="openstack-operators/openstack-operator-index-lgzqc" Mar 20 09:13:26 crc kubenswrapper[4823]: I0320 09:13:26.025476 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk9wh\" (UniqueName: \"kubernetes.io/projected/75433774-57f1-47f6-9472-d36aa486ea2e-kube-api-access-hk9wh\") pod \"openstack-operator-index-lgzqc\" (UID: \"75433774-57f1-47f6-9472-d36aa486ea2e\") " pod="openstack-operators/openstack-operator-index-lgzqc" Mar 20 09:13:26 crc kubenswrapper[4823]: I0320 09:13:26.124360 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lgzqc" Mar 20 09:13:26 crc kubenswrapper[4823]: I0320 09:13:26.458377 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:26 crc kubenswrapper[4823]: I0320 09:13:26.529634 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:26 crc kubenswrapper[4823]: I0320 09:13:26.583880 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lgzqc"] Mar 20 09:13:26 crc kubenswrapper[4823]: W0320 09:13:26.584032 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75433774_57f1_47f6_9472_d36aa486ea2e.slice/crio-ea61fd54f22f69d1580a0c57d6e0e073109d40e1e87135e223fcd6e974efa930 WatchSource:0}: Error finding container ea61fd54f22f69d1580a0c57d6e0e073109d40e1e87135e223fcd6e974efa930: Status 404 returned error can't find the container with id ea61fd54f22f69d1580a0c57d6e0e073109d40e1e87135e223fcd6e974efa930 Mar 20 09:13:26 crc kubenswrapper[4823]: I0320 09:13:26.974987 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lgzqc" event={"ID":"75433774-57f1-47f6-9472-d36aa486ea2e","Type":"ContainerStarted","Data":"ea61fd54f22f69d1580a0c57d6e0e073109d40e1e87135e223fcd6e974efa930"} Mar 20 09:13:28 crc kubenswrapper[4823]: I0320 09:13:28.965839 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-lgzqc"] Mar 20 09:13:29 crc kubenswrapper[4823]: I0320 09:13:29.572319 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-55sh4"] Mar 20 09:13:29 crc kubenswrapper[4823]: I0320 09:13:29.573362 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-55sh4" Mar 20 09:13:29 crc kubenswrapper[4823]: I0320 09:13:29.582164 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-55sh4"] Mar 20 09:13:29 crc kubenswrapper[4823]: I0320 09:13:29.662247 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxgtk\" (UniqueName: \"kubernetes.io/projected/735b38f1-4ef0-411b-97a1-f48fef038a2f-kube-api-access-zxgtk\") pod \"openstack-operator-index-55sh4\" (UID: \"735b38f1-4ef0-411b-97a1-f48fef038a2f\") " pod="openstack-operators/openstack-operator-index-55sh4" Mar 20 09:13:29 crc kubenswrapper[4823]: I0320 09:13:29.764861 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxgtk\" (UniqueName: \"kubernetes.io/projected/735b38f1-4ef0-411b-97a1-f48fef038a2f-kube-api-access-zxgtk\") pod \"openstack-operator-index-55sh4\" (UID: \"735b38f1-4ef0-411b-97a1-f48fef038a2f\") " pod="openstack-operators/openstack-operator-index-55sh4" Mar 20 09:13:29 crc kubenswrapper[4823]: I0320 09:13:29.797413 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxgtk\" (UniqueName: \"kubernetes.io/projected/735b38f1-4ef0-411b-97a1-f48fef038a2f-kube-api-access-zxgtk\") pod \"openstack-operator-index-55sh4\" (UID: \"735b38f1-4ef0-411b-97a1-f48fef038a2f\") " pod="openstack-operators/openstack-operator-index-55sh4" Mar 20 09:13:29 crc kubenswrapper[4823]: I0320 09:13:29.890366 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-55sh4" Mar 20 09:13:30 crc kubenswrapper[4823]: I0320 09:13:30.012975 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lgzqc" event={"ID":"75433774-57f1-47f6-9472-d36aa486ea2e","Type":"ContainerStarted","Data":"deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f"} Mar 20 09:13:30 crc kubenswrapper[4823]: I0320 09:13:30.013104 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-lgzqc" podUID="75433774-57f1-47f6-9472-d36aa486ea2e" containerName="registry-server" containerID="cri-o://deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f" gracePeriod=2 Mar 20 09:13:30 crc kubenswrapper[4823]: I0320 09:13:30.031527 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-lgzqc" podStartSLOduration=2.213997452 podStartE2EDuration="5.031509204s" podCreationTimestamp="2026-03-20 09:13:25 +0000 UTC" firstStartedPulling="2026-03-20 09:13:26.586939643 +0000 UTC m=+973.272836017" lastFinishedPulling="2026-03-20 09:13:29.404451395 +0000 UTC m=+976.090347769" observedRunningTime="2026-03-20 09:13:30.02883328 +0000 UTC m=+976.714729654" watchObservedRunningTime="2026-03-20 09:13:30.031509204 +0000 UTC m=+976.717405578" Mar 20 09:13:30 crc kubenswrapper[4823]: I0320 09:13:30.312385 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-55sh4"] Mar 20 09:13:30 crc kubenswrapper[4823]: I0320 09:13:30.360653 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lgzqc" Mar 20 09:13:30 crc kubenswrapper[4823]: I0320 09:13:30.374363 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk9wh\" (UniqueName: \"kubernetes.io/projected/75433774-57f1-47f6-9472-d36aa486ea2e-kube-api-access-hk9wh\") pod \"75433774-57f1-47f6-9472-d36aa486ea2e\" (UID: \"75433774-57f1-47f6-9472-d36aa486ea2e\") " Mar 20 09:13:30 crc kubenswrapper[4823]: I0320 09:13:30.379081 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75433774-57f1-47f6-9472-d36aa486ea2e-kube-api-access-hk9wh" (OuterVolumeSpecName: "kube-api-access-hk9wh") pod "75433774-57f1-47f6-9472-d36aa486ea2e" (UID: "75433774-57f1-47f6-9472-d36aa486ea2e"). InnerVolumeSpecName "kube-api-access-hk9wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:13:30 crc kubenswrapper[4823]: I0320 09:13:30.475826 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk9wh\" (UniqueName: \"kubernetes.io/projected/75433774-57f1-47f6-9472-d36aa486ea2e-kube-api-access-hk9wh\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.023603 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-55sh4" event={"ID":"735b38f1-4ef0-411b-97a1-f48fef038a2f","Type":"ContainerStarted","Data":"2abefff2ac61b03c4075e1727ee12701fce6427588bb5ff16b5da9ed2ae7bccb"} Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.023987 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-55sh4" event={"ID":"735b38f1-4ef0-411b-97a1-f48fef038a2f","Type":"ContainerStarted","Data":"1255deb18b73a786298654323f3bc9c093ea50b861e20051f7318f1c5efd6a19"} Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.027526 4823 generic.go:334] "Generic (PLEG): container finished" podID="75433774-57f1-47f6-9472-d36aa486ea2e" containerID="deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f" exitCode=0 Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.027660 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lgzqc" Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.027679 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lgzqc" event={"ID":"75433774-57f1-47f6-9472-d36aa486ea2e","Type":"ContainerDied","Data":"deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f"} Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.027860 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lgzqc" event={"ID":"75433774-57f1-47f6-9472-d36aa486ea2e","Type":"ContainerDied","Data":"ea61fd54f22f69d1580a0c57d6e0e073109d40e1e87135e223fcd6e974efa930"} Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.027933 4823 scope.go:117] "RemoveContainer" containerID="deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f" Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.047269 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-55sh4" podStartSLOduration=2.005149881 podStartE2EDuration="2.047248471s" podCreationTimestamp="2026-03-20 09:13:29 +0000 UTC" firstStartedPulling="2026-03-20 09:13:30.323471362 +0000 UTC m=+977.009367746" lastFinishedPulling="2026-03-20 09:13:30.365569962 +0000 UTC m=+977.051466336" observedRunningTime="2026-03-20 09:13:31.043966509 +0000 UTC m=+977.729862923" watchObservedRunningTime="2026-03-20 09:13:31.047248471 +0000 UTC m=+977.733144845" Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.059450 4823 scope.go:117] "RemoveContainer" containerID="deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f" Mar 20 09:13:31 crc kubenswrapper[4823]: E0320 09:13:31.059814 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f\": container with ID starting with deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f not found: ID does not exist" containerID="deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f" Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.059853 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f"} err="failed to get container status \"deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f\": rpc error: code = NotFound desc = could not find container \"deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f\": container with ID starting with deb37e6f93d63217d67a99585d9a8f9d8e8292e1d98ac39dcb4db1922708d37f not found: ID does not exist" Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.065242 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-lgzqc"] Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.069728 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-lgzqc"] Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.461523 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-zgtrw" Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.476815 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-trvvd" Mar 20 09:13:31 crc kubenswrapper[4823]: I0320 09:13:31.548039 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-jh928" Mar 20 09:13:32 crc kubenswrapper[4823]: I0320 09:13:32.105789 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75433774-57f1-47f6-9472-d36aa486ea2e" path="/var/lib/kubelet/pods/75433774-57f1-47f6-9472-d36aa486ea2e/volumes" Mar 20 09:13:39 crc kubenswrapper[4823]: I0320 09:13:39.430219 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:13:39 crc kubenswrapper[4823]: I0320 09:13:39.430854 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:13:39 crc kubenswrapper[4823]: I0320 09:13:39.891344 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-55sh4" Mar 20 09:13:39 crc kubenswrapper[4823]: I0320 09:13:39.891444 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-55sh4" Mar 20 09:13:39 crc kubenswrapper[4823]: I0320 09:13:39.932558 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-55sh4" Mar 20 09:13:40 crc kubenswrapper[4823]: I0320 09:13:40.122674 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-55sh4" Mar 20 09:13:41 crc kubenswrapper[4823]: I0320 09:13:41.874292 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4"] Mar 20 09:13:41 crc kubenswrapper[4823]: E0320 09:13:41.874845 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75433774-57f1-47f6-9472-d36aa486ea2e" containerName="registry-server" Mar 20 09:13:41 crc kubenswrapper[4823]: I0320 09:13:41.874865 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="75433774-57f1-47f6-9472-d36aa486ea2e" containerName="registry-server" Mar 20 09:13:41 crc kubenswrapper[4823]: I0320 09:13:41.874997 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="75433774-57f1-47f6-9472-d36aa486ea2e" containerName="registry-server" Mar 20 09:13:41 crc kubenswrapper[4823]: I0320 09:13:41.875978 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:41 crc kubenswrapper[4823]: I0320 09:13:41.877973 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-vfg8x" Mar 20 09:13:41 crc kubenswrapper[4823]: I0320 09:13:41.884257 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4"] Mar 20 09:13:41 crc kubenswrapper[4823]: I0320 09:13:41.906222 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq8xj\" (UniqueName: \"kubernetes.io/projected/627dd666-3cdd-4c47-939c-9061526fac3d-kube-api-access-mq8xj\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:41 crc kubenswrapper[4823]: I0320 09:13:41.906485 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-util\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:41 crc kubenswrapper[4823]: I0320 09:13:41.906642 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-bundle\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:42 crc kubenswrapper[4823]: I0320 09:13:42.007219 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq8xj\" (UniqueName: \"kubernetes.io/projected/627dd666-3cdd-4c47-939c-9061526fac3d-kube-api-access-mq8xj\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:42 crc kubenswrapper[4823]: I0320 09:13:42.007270 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-util\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:42 crc kubenswrapper[4823]: I0320 09:13:42.007312 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-bundle\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:42 crc kubenswrapper[4823]: I0320 09:13:42.007719 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-bundle\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:42 crc kubenswrapper[4823]: I0320 09:13:42.008162 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-util\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:42 crc kubenswrapper[4823]: I0320 09:13:42.031057 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq8xj\" (UniqueName: \"kubernetes.io/projected/627dd666-3cdd-4c47-939c-9061526fac3d-kube-api-access-mq8xj\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:42 crc kubenswrapper[4823]: I0320 09:13:42.209430 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:42 crc kubenswrapper[4823]: W0320 09:13:42.656142 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod627dd666_3cdd_4c47_939c_9061526fac3d.slice/crio-f4018c69cc9099d474ad9eb7c2b1a234f69d14f7a27350936feb4067d3252223 WatchSource:0}: Error finding container f4018c69cc9099d474ad9eb7c2b1a234f69d14f7a27350936feb4067d3252223: Status 404 returned error can't find the container with id f4018c69cc9099d474ad9eb7c2b1a234f69d14f7a27350936feb4067d3252223 Mar 20 09:13:42 crc kubenswrapper[4823]: I0320 09:13:42.658732 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4"] Mar 20 09:13:43 crc kubenswrapper[4823]: I0320 09:13:43.115355 4823 generic.go:334] "Generic (PLEG): container finished" podID="627dd666-3cdd-4c47-939c-9061526fac3d" containerID="130f022cad81dc5b3db55444f19a2ae36b3e75510ef3226b7ee89c31e95d7b39" exitCode=0 Mar 20 09:13:43 crc kubenswrapper[4823]: I0320 09:13:43.115407 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" event={"ID":"627dd666-3cdd-4c47-939c-9061526fac3d","Type":"ContainerDied","Data":"130f022cad81dc5b3db55444f19a2ae36b3e75510ef3226b7ee89c31e95d7b39"} Mar 20 09:13:43 crc kubenswrapper[4823]: I0320 09:13:43.115433 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" event={"ID":"627dd666-3cdd-4c47-939c-9061526fac3d","Type":"ContainerStarted","Data":"f4018c69cc9099d474ad9eb7c2b1a234f69d14f7a27350936feb4067d3252223"} Mar 20 09:13:44 crc kubenswrapper[4823]: I0320 09:13:44.121225 4823 generic.go:334] "Generic (PLEG): container finished" podID="627dd666-3cdd-4c47-939c-9061526fac3d" containerID="4eca0aa1744e0be56d71edfd165e0d015c027db855712d6629701036023d65a9" exitCode=0 Mar 20 09:13:44 crc kubenswrapper[4823]: I0320 09:13:44.121263 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" event={"ID":"627dd666-3cdd-4c47-939c-9061526fac3d","Type":"ContainerDied","Data":"4eca0aa1744e0be56d71edfd165e0d015c027db855712d6629701036023d65a9"} Mar 20 09:13:45 crc kubenswrapper[4823]: I0320 09:13:45.129597 4823 generic.go:334] "Generic (PLEG): container finished" podID="627dd666-3cdd-4c47-939c-9061526fac3d" containerID="9b95f011f1be0e16beb5aa7c313ba0a9d3a1dce1bca25e3e4d781404080013b7" exitCode=0 Mar 20 09:13:45 crc kubenswrapper[4823]: I0320 09:13:45.129638 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" event={"ID":"627dd666-3cdd-4c47-939c-9061526fac3d","Type":"ContainerDied","Data":"9b95f011f1be0e16beb5aa7c313ba0a9d3a1dce1bca25e3e4d781404080013b7"} Mar 20 09:13:46 crc kubenswrapper[4823]: I0320 09:13:46.429415 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:46 crc kubenswrapper[4823]: I0320 09:13:46.560892 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq8xj\" (UniqueName: \"kubernetes.io/projected/627dd666-3cdd-4c47-939c-9061526fac3d-kube-api-access-mq8xj\") pod \"627dd666-3cdd-4c47-939c-9061526fac3d\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " Mar 20 09:13:46 crc kubenswrapper[4823]: I0320 09:13:46.560968 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-bundle\") pod \"627dd666-3cdd-4c47-939c-9061526fac3d\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " Mar 20 09:13:46 crc kubenswrapper[4823]: I0320 09:13:46.561041 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-util\") pod \"627dd666-3cdd-4c47-939c-9061526fac3d\" (UID: \"627dd666-3cdd-4c47-939c-9061526fac3d\") " Mar 20 09:13:46 crc kubenswrapper[4823]: I0320 09:13:46.561667 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-bundle" (OuterVolumeSpecName: "bundle") pod "627dd666-3cdd-4c47-939c-9061526fac3d" (UID: "627dd666-3cdd-4c47-939c-9061526fac3d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:13:46 crc kubenswrapper[4823]: I0320 09:13:46.566941 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/627dd666-3cdd-4c47-939c-9061526fac3d-kube-api-access-mq8xj" (OuterVolumeSpecName: "kube-api-access-mq8xj") pod "627dd666-3cdd-4c47-939c-9061526fac3d" (UID: "627dd666-3cdd-4c47-939c-9061526fac3d"). InnerVolumeSpecName "kube-api-access-mq8xj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:13:46 crc kubenswrapper[4823]: I0320 09:13:46.577339 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-util" (OuterVolumeSpecName: "util") pod "627dd666-3cdd-4c47-939c-9061526fac3d" (UID: "627dd666-3cdd-4c47-939c-9061526fac3d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:13:46 crc kubenswrapper[4823]: I0320 09:13:46.662888 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq8xj\" (UniqueName: \"kubernetes.io/projected/627dd666-3cdd-4c47-939c-9061526fac3d-kube-api-access-mq8xj\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:46 crc kubenswrapper[4823]: I0320 09:13:46.662923 4823 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:46 crc kubenswrapper[4823]: I0320 09:13:46.662932 4823 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/627dd666-3cdd-4c47-939c-9061526fac3d-util\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:47 crc kubenswrapper[4823]: I0320 09:13:47.154321 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" event={"ID":"627dd666-3cdd-4c47-939c-9061526fac3d","Type":"ContainerDied","Data":"f4018c69cc9099d474ad9eb7c2b1a234f69d14f7a27350936feb4067d3252223"} Mar 20 09:13:47 crc kubenswrapper[4823]: I0320 09:13:47.154364 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4018c69cc9099d474ad9eb7c2b1a234f69d14f7a27350936feb4067d3252223" Mar 20 09:13:47 crc kubenswrapper[4823]: I0320 09:13:47.154422 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4" Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.422710 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr"] Mar 20 09:13:49 crc kubenswrapper[4823]: E0320 09:13:49.423020 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627dd666-3cdd-4c47-939c-9061526fac3d" containerName="extract" Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.423035 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="627dd666-3cdd-4c47-939c-9061526fac3d" containerName="extract" Mar 20 09:13:49 crc kubenswrapper[4823]: E0320 09:13:49.423058 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627dd666-3cdd-4c47-939c-9061526fac3d" containerName="pull" Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.423066 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="627dd666-3cdd-4c47-939c-9061526fac3d" containerName="pull" Mar 20 09:13:49 crc kubenswrapper[4823]: E0320 09:13:49.423080 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627dd666-3cdd-4c47-939c-9061526fac3d" containerName="util" Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.423088 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="627dd666-3cdd-4c47-939c-9061526fac3d" containerName="util" Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.423227 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="627dd666-3cdd-4c47-939c-9061526fac3d" containerName="extract" Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.423681 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr" Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.427136 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-ffsv7" Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.468810 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr"] Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.520197 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2fd7\" (UniqueName: \"kubernetes.io/projected/328e53ac-f20f-4a71-b756-f143cd0fc077-kube-api-access-s2fd7\") pod \"openstack-operator-controller-init-9df8dd5fd-jxnmr\" (UID: \"328e53ac-f20f-4a71-b756-f143cd0fc077\") " pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr" Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.621621 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2fd7\" (UniqueName: \"kubernetes.io/projected/328e53ac-f20f-4a71-b756-f143cd0fc077-kube-api-access-s2fd7\") pod \"openstack-operator-controller-init-9df8dd5fd-jxnmr\" (UID: \"328e53ac-f20f-4a71-b756-f143cd0fc077\") " pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr" Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.642525 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2fd7\" (UniqueName: \"kubernetes.io/projected/328e53ac-f20f-4a71-b756-f143cd0fc077-kube-api-access-s2fd7\") pod \"openstack-operator-controller-init-9df8dd5fd-jxnmr\" (UID: \"328e53ac-f20f-4a71-b756-f143cd0fc077\") " pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr" Mar 20 09:13:49 crc kubenswrapper[4823]: I0320 09:13:49.738151 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr" Mar 20 09:13:50 crc kubenswrapper[4823]: I0320 09:13:50.773310 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr"] Mar 20 09:13:50 crc kubenswrapper[4823]: W0320 09:13:50.802796 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod328e53ac_f20f_4a71_b756_f143cd0fc077.slice/crio-6c1674ef134f389c3793973c999d37de4057d8458b7d9d8140dedc5807c21a66 WatchSource:0}: Error finding container 6c1674ef134f389c3793973c999d37de4057d8458b7d9d8140dedc5807c21a66: Status 404 returned error can't find the container with id 6c1674ef134f389c3793973c999d37de4057d8458b7d9d8140dedc5807c21a66 Mar 20 09:13:51 crc kubenswrapper[4823]: I0320 09:13:51.177204 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr" event={"ID":"328e53ac-f20f-4a71-b756-f143cd0fc077","Type":"ContainerStarted","Data":"6c1674ef134f389c3793973c999d37de4057d8458b7d9d8140dedc5807c21a66"} Mar 20 09:13:55 crc kubenswrapper[4823]: I0320 09:13:55.210065 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr" event={"ID":"328e53ac-f20f-4a71-b756-f143cd0fc077","Type":"ContainerStarted","Data":"02ab17d09ef941cc5d22f87321baf9878b20dbb1c533d6e4620e0b1c5b59e3f4"} Mar 20 09:13:55 crc kubenswrapper[4823]: I0320 09:13:55.211165 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr" Mar 20 09:13:55 crc kubenswrapper[4823]: I0320 09:13:55.247594 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr" podStartSLOduration=2.657474094 podStartE2EDuration="6.247575643s" podCreationTimestamp="2026-03-20 09:13:49 +0000 UTC" firstStartedPulling="2026-03-20 09:13:50.804932656 +0000 UTC m=+997.490829040" lastFinishedPulling="2026-03-20 09:13:54.395034215 +0000 UTC m=+1001.080930589" observedRunningTime="2026-03-20 09:13:55.245416516 +0000 UTC m=+1001.931312950" watchObservedRunningTime="2026-03-20 09:13:55.247575643 +0000 UTC m=+1001.933472027" Mar 20 09:13:59 crc kubenswrapper[4823]: I0320 09:13:59.740638 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-jxnmr" Mar 20 09:14:00 crc kubenswrapper[4823]: I0320 09:14:00.130487 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566634-78xrx"] Mar 20 09:14:00 crc kubenswrapper[4823]: I0320 09:14:00.131776 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566634-78xrx" Mar 20 09:14:00 crc kubenswrapper[4823]: I0320 09:14:00.136382 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:14:00 crc kubenswrapper[4823]: I0320 09:14:00.136833 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:14:00 crc kubenswrapper[4823]: I0320 09:14:00.137499 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:14:00 crc kubenswrapper[4823]: I0320 09:14:00.154502 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566634-78xrx"] Mar 20 09:14:00 crc kubenswrapper[4823]: I0320 09:14:00.171287 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfn94\" (UniqueName: \"kubernetes.io/projected/37e5ce93-ad64-4864-9430-45fb877fd038-kube-api-access-tfn94\") pod \"auto-csr-approver-29566634-78xrx\" (UID: \"37e5ce93-ad64-4864-9430-45fb877fd038\") " pod="openshift-infra/auto-csr-approver-29566634-78xrx" Mar 20 09:14:00 crc kubenswrapper[4823]: I0320 09:14:00.272295 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfn94\" (UniqueName: \"kubernetes.io/projected/37e5ce93-ad64-4864-9430-45fb877fd038-kube-api-access-tfn94\") pod \"auto-csr-approver-29566634-78xrx\" (UID: \"37e5ce93-ad64-4864-9430-45fb877fd038\") " pod="openshift-infra/auto-csr-approver-29566634-78xrx" Mar 20 09:14:00 crc kubenswrapper[4823]: I0320 09:14:00.289281 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfn94\" (UniqueName: \"kubernetes.io/projected/37e5ce93-ad64-4864-9430-45fb877fd038-kube-api-access-tfn94\") pod \"auto-csr-approver-29566634-78xrx\" (UID: \"37e5ce93-ad64-4864-9430-45fb877fd038\") " pod="openshift-infra/auto-csr-approver-29566634-78xrx" Mar 20 09:14:00 crc kubenswrapper[4823]: I0320 09:14:00.466441 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566634-78xrx" Mar 20 09:14:00 crc kubenswrapper[4823]: I0320 09:14:00.972523 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566634-78xrx"] Mar 20 09:14:01 crc kubenswrapper[4823]: I0320 09:14:01.249476 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566634-78xrx" event={"ID":"37e5ce93-ad64-4864-9430-45fb877fd038","Type":"ContainerStarted","Data":"2c151987859c1814441bd1ac574c4bc129fbb1ce8bcd1f91c012bcf75451348b"} Mar 20 09:14:03 crc kubenswrapper[4823]: I0320 09:14:03.264015 4823 generic.go:334] "Generic (PLEG): container finished" podID="37e5ce93-ad64-4864-9430-45fb877fd038" containerID="4378b83304dcb3dd2ae8300b75d65d9c7f24a2a9af713826305396acafa0651d" exitCode=0 Mar 20 09:14:03 crc kubenswrapper[4823]: I0320 09:14:03.264080 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566634-78xrx" event={"ID":"37e5ce93-ad64-4864-9430-45fb877fd038","Type":"ContainerDied","Data":"4378b83304dcb3dd2ae8300b75d65d9c7f24a2a9af713826305396acafa0651d"} Mar 20 09:14:04 crc kubenswrapper[4823]: I0320 09:14:04.547806 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566634-78xrx" Mar 20 09:14:04 crc kubenswrapper[4823]: I0320 09:14:04.729827 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfn94\" (UniqueName: \"kubernetes.io/projected/37e5ce93-ad64-4864-9430-45fb877fd038-kube-api-access-tfn94\") pod \"37e5ce93-ad64-4864-9430-45fb877fd038\" (UID: \"37e5ce93-ad64-4864-9430-45fb877fd038\") " Mar 20 09:14:04 crc kubenswrapper[4823]: I0320 09:14:04.746355 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37e5ce93-ad64-4864-9430-45fb877fd038-kube-api-access-tfn94" (OuterVolumeSpecName: "kube-api-access-tfn94") pod "37e5ce93-ad64-4864-9430-45fb877fd038" (UID: "37e5ce93-ad64-4864-9430-45fb877fd038"). InnerVolumeSpecName "kube-api-access-tfn94". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:14:04 crc kubenswrapper[4823]: I0320 09:14:04.831806 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfn94\" (UniqueName: \"kubernetes.io/projected/37e5ce93-ad64-4864-9430-45fb877fd038-kube-api-access-tfn94\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:05 crc kubenswrapper[4823]: I0320 09:14:05.276809 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566634-78xrx" event={"ID":"37e5ce93-ad64-4864-9430-45fb877fd038","Type":"ContainerDied","Data":"2c151987859c1814441bd1ac574c4bc129fbb1ce8bcd1f91c012bcf75451348b"} Mar 20 09:14:05 crc kubenswrapper[4823]: I0320 09:14:05.276855 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c151987859c1814441bd1ac574c4bc129fbb1ce8bcd1f91c012bcf75451348b" Mar 20 09:14:05 crc kubenswrapper[4823]: I0320 09:14:05.276864 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566634-78xrx" Mar 20 09:14:05 crc kubenswrapper[4823]: I0320 09:14:05.659066 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566628-gq4lq"] Mar 20 09:14:05 crc kubenswrapper[4823]: I0320 09:14:05.664334 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566628-gq4lq"] Mar 20 09:14:06 crc kubenswrapper[4823]: I0320 09:14:06.105427 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dc8b1fb-cbe3-4e67-8c08-878ac01d1853" path="/var/lib/kubelet/pods/7dc8b1fb-cbe3-4e67-8c08-878ac01d1853/volumes" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.540828 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mdc24"] Mar 20 09:14:07 crc kubenswrapper[4823]: E0320 09:14:07.541309 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37e5ce93-ad64-4864-9430-45fb877fd038" containerName="oc" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.541321 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="37e5ce93-ad64-4864-9430-45fb877fd038" containerName="oc" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.541425 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="37e5ce93-ad64-4864-9430-45fb877fd038" containerName="oc" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.542285 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.560583 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdc24"] Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.694954 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2kcb\" (UniqueName: \"kubernetes.io/projected/d1ce5c2e-da8c-4004-ab2a-790f563accfb-kube-api-access-b2kcb\") pod \"redhat-marketplace-mdc24\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.695050 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-catalog-content\") pod \"redhat-marketplace-mdc24\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.695078 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-utilities\") pod \"redhat-marketplace-mdc24\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.796672 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2kcb\" (UniqueName: \"kubernetes.io/projected/d1ce5c2e-da8c-4004-ab2a-790f563accfb-kube-api-access-b2kcb\") pod \"redhat-marketplace-mdc24\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.796798 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-catalog-content\") pod \"redhat-marketplace-mdc24\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.796847 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-utilities\") pod \"redhat-marketplace-mdc24\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.797320 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-catalog-content\") pod \"redhat-marketplace-mdc24\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.797379 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-utilities\") pod \"redhat-marketplace-mdc24\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.818446 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2kcb\" (UniqueName: \"kubernetes.io/projected/d1ce5c2e-da8c-4004-ab2a-790f563accfb-kube-api-access-b2kcb\") pod \"redhat-marketplace-mdc24\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:07 crc kubenswrapper[4823]: I0320 09:14:07.855323 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:08 crc kubenswrapper[4823]: I0320 09:14:08.254402 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdc24"] Mar 20 09:14:08 crc kubenswrapper[4823]: W0320 09:14:08.261756 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1ce5c2e_da8c_4004_ab2a_790f563accfb.slice/crio-122f00e5a9f4537a184457b34f3b4f57b751aee4fd46440de06a34ed98e338f4 WatchSource:0}: Error finding container 122f00e5a9f4537a184457b34f3b4f57b751aee4fd46440de06a34ed98e338f4: Status 404 returned error can't find the container with id 122f00e5a9f4537a184457b34f3b4f57b751aee4fd46440de06a34ed98e338f4 Mar 20 09:14:08 crc kubenswrapper[4823]: I0320 09:14:08.294624 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdc24" event={"ID":"d1ce5c2e-da8c-4004-ab2a-790f563accfb","Type":"ContainerStarted","Data":"122f00e5a9f4537a184457b34f3b4f57b751aee4fd46440de06a34ed98e338f4"} Mar 20 09:14:09 crc kubenswrapper[4823]: I0320 09:14:09.301259 4823 generic.go:334] "Generic (PLEG): container finished" podID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" containerID="47d0af9108752feb6de2614ae3ef6b312f4709c87efc10b180c1a0b2bdc3f167" exitCode=0 Mar 20 09:14:09 crc kubenswrapper[4823]: I0320 09:14:09.301310 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdc24" event={"ID":"d1ce5c2e-da8c-4004-ab2a-790f563accfb","Type":"ContainerDied","Data":"47d0af9108752feb6de2614ae3ef6b312f4709c87efc10b180c1a0b2bdc3f167"} Mar 20 09:14:09 crc kubenswrapper[4823]: I0320 09:14:09.430253 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:14:09 crc kubenswrapper[4823]: I0320 09:14:09.430308 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:14:09 crc kubenswrapper[4823]: I0320 09:14:09.430350 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:14:09 crc kubenswrapper[4823]: I0320 09:14:09.430933 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c6623b862df4196951c10211c83ea8d06f08a5d6875319545662d0ea4663061"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:14:09 crc kubenswrapper[4823]: I0320 09:14:09.431003 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://3c6623b862df4196951c10211c83ea8d06f08a5d6875319545662d0ea4663061" gracePeriod=600 Mar 20 09:14:10 crc kubenswrapper[4823]: I0320 09:14:10.311047 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="3c6623b862df4196951c10211c83ea8d06f08a5d6875319545662d0ea4663061" exitCode=0 Mar 20 09:14:10 crc kubenswrapper[4823]: I0320 09:14:10.311114 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"3c6623b862df4196951c10211c83ea8d06f08a5d6875319545662d0ea4663061"} Mar 20 09:14:10 crc kubenswrapper[4823]: I0320 09:14:10.312704 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"0b8674863c83970000faa79c02318a604d3777e9ca2e37590494e528f8d2cbc7"} Mar 20 09:14:10 crc kubenswrapper[4823]: I0320 09:14:10.312778 4823 scope.go:117] "RemoveContainer" containerID="60ec5c90c519363b5a41f87bdb8567fcf49b737ce1cfeea150ceae7849bf0158" Mar 20 09:14:11 crc kubenswrapper[4823]: I0320 09:14:11.322968 4823 generic.go:334] "Generic (PLEG): container finished" podID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" containerID="933dc4e57a873cda6b684cdc8d732e377a16cf8c796fe00e1807c57227303c1f" exitCode=0 Mar 20 09:14:11 crc kubenswrapper[4823]: I0320 09:14:11.323100 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdc24" event={"ID":"d1ce5c2e-da8c-4004-ab2a-790f563accfb","Type":"ContainerDied","Data":"933dc4e57a873cda6b684cdc8d732e377a16cf8c796fe00e1807c57227303c1f"} Mar 20 09:14:12 crc kubenswrapper[4823]: I0320 09:14:12.339556 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdc24" event={"ID":"d1ce5c2e-da8c-4004-ab2a-790f563accfb","Type":"ContainerStarted","Data":"823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902"} Mar 20 09:14:12 crc kubenswrapper[4823]: I0320 09:14:12.359824 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mdc24" podStartSLOduration=2.754551976 podStartE2EDuration="5.35980863s" podCreationTimestamp="2026-03-20 09:14:07 +0000 UTC" firstStartedPulling="2026-03-20 09:14:09.303858908 +0000 UTC m=+1015.989755282" lastFinishedPulling="2026-03-20 09:14:11.909115562 +0000 UTC m=+1018.595011936" observedRunningTime="2026-03-20 09:14:12.357398225 +0000 UTC m=+1019.043294599" watchObservedRunningTime="2026-03-20 09:14:12.35980863 +0000 UTC m=+1019.045704994" Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.563444 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-92fnv"] Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.565266 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.634600 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-92fnv"] Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.757588 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-utilities\") pod \"community-operators-92fnv\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.757776 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-catalog-content\") pod \"community-operators-92fnv\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.757819 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j9gg\" (UniqueName: \"kubernetes.io/projected/57e15159-c2f8-45a9-b37a-2ed8529d73c7-kube-api-access-4j9gg\") pod \"community-operators-92fnv\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.858637 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-catalog-content\") pod \"community-operators-92fnv\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.858696 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j9gg\" (UniqueName: \"kubernetes.io/projected/57e15159-c2f8-45a9-b37a-2ed8529d73c7-kube-api-access-4j9gg\") pod \"community-operators-92fnv\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.858731 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-utilities\") pod \"community-operators-92fnv\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.859330 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-utilities\") pod \"community-operators-92fnv\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.859422 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-catalog-content\") pod \"community-operators-92fnv\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.878264 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j9gg\" (UniqueName: \"kubernetes.io/projected/57e15159-c2f8-45a9-b37a-2ed8529d73c7-kube-api-access-4j9gg\") pod \"community-operators-92fnv\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:15 crc kubenswrapper[4823]: I0320 09:14:15.881923 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:16 crc kubenswrapper[4823]: I0320 09:14:16.286025 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-92fnv"] Mar 20 09:14:16 crc kubenswrapper[4823]: W0320 09:14:16.287597 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57e15159_c2f8_45a9_b37a_2ed8529d73c7.slice/crio-c0eb1d25654d9de886a0cfea9f32bcb123bc86c80eb0273586dee5d06574ba36 WatchSource:0}: Error finding container c0eb1d25654d9de886a0cfea9f32bcb123bc86c80eb0273586dee5d06574ba36: Status 404 returned error can't find the container with id c0eb1d25654d9de886a0cfea9f32bcb123bc86c80eb0273586dee5d06574ba36 Mar 20 09:14:16 crc kubenswrapper[4823]: I0320 09:14:16.377589 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92fnv" event={"ID":"57e15159-c2f8-45a9-b37a-2ed8529d73c7","Type":"ContainerStarted","Data":"c0eb1d25654d9de886a0cfea9f32bcb123bc86c80eb0273586dee5d06574ba36"} Mar 20 09:14:17 crc kubenswrapper[4823]: I0320 09:14:17.396119 4823 generic.go:334] "Generic (PLEG): container finished" podID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerID="3a1793ae11f311dc48be360d6e1b3fda6998de43b9b18d263073ae39f8d7f6b5" exitCode=0 Mar 20 09:14:17 crc kubenswrapper[4823]: I0320 09:14:17.398998 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92fnv" event={"ID":"57e15159-c2f8-45a9-b37a-2ed8529d73c7","Type":"ContainerDied","Data":"3a1793ae11f311dc48be360d6e1b3fda6998de43b9b18d263073ae39f8d7f6b5"} Mar 20 09:14:17 crc kubenswrapper[4823]: I0320 09:14:17.856407 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:17 crc kubenswrapper[4823]: I0320 09:14:17.856467 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:17 crc kubenswrapper[4823]: I0320 09:14:17.914583 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:17 crc kubenswrapper[4823]: I0320 09:14:17.984121 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx"] Mar 20 09:14:17 crc kubenswrapper[4823]: I0320 09:14:17.984810 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx" Mar 20 09:14:17 crc kubenswrapper[4823]: I0320 09:14:17.994719 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2"] Mar 20 09:14:17 crc kubenswrapper[4823]: I0320 09:14:17.996252 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2" Mar 20 09:14:17 crc kubenswrapper[4823]: I0320 09:14:17.996338 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qhfzb" Mar 20 09:14:17 crc kubenswrapper[4823]: I0320 09:14:17.998918 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-6tzhs" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.012449 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.024362 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.031403 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.032363 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.034363 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-p27qb" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.057798 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.058786 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.067825 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.068308 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-kz5fw" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.079273 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.105766 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8hf6\" (UniqueName: \"kubernetes.io/projected/f99021be-4dee-4443-938e-ec54f85d61f5-kube-api-access-x8hf6\") pod \"designate-operator-controller-manager-588d4d986b-56xqd\" (UID: \"f99021be-4dee-4443-938e-ec54f85d61f5\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.105848 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l2z6\" (UniqueName: \"kubernetes.io/projected/a95e604d-a508-41bc-84e3-78fd32a90ed4-kube-api-access-6l2z6\") pod \"cinder-operator-controller-manager-8d58dc466-47nf2\" (UID: \"a95e604d-a508-41bc-84e3-78fd32a90ed4\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.105898 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4jl6\" (UniqueName: \"kubernetes.io/projected/c9b3eb51-1ddc-4abc-9d8a-09611ef3983c-kube-api-access-r4jl6\") pod \"barbican-operator-controller-manager-59bc569d95-z9bkx\" (UID: \"c9b3eb51-1ddc-4abc-9d8a-09611ef3983c\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.105957 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tvg8\" (UniqueName: \"kubernetes.io/projected/70a65110-b953-43ca-a6ec-5131b4660aea-kube-api-access-5tvg8\") pod \"glance-operator-controller-manager-79df6bcc97-7c7hb\" (UID: \"70a65110-b953-43ca-a6ec-5131b4660aea\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.115158 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.115822 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.118800 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-kknks" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.148970 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.150082 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.153993 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-bkbbg" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.167823 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.190602 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.213778 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tvg8\" (UniqueName: \"kubernetes.io/projected/70a65110-b953-43ca-a6ec-5131b4660aea-kube-api-access-5tvg8\") pod \"glance-operator-controller-manager-79df6bcc97-7c7hb\" (UID: \"70a65110-b953-43ca-a6ec-5131b4660aea\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.214106 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8hf6\" (UniqueName: \"kubernetes.io/projected/f99021be-4dee-4443-938e-ec54f85d61f5-kube-api-access-x8hf6\") pod \"designate-operator-controller-manager-588d4d986b-56xqd\" (UID: \"f99021be-4dee-4443-938e-ec54f85d61f5\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.214173 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l2z6\" (UniqueName: \"kubernetes.io/projected/a95e604d-a508-41bc-84e3-78fd32a90ed4-kube-api-access-6l2z6\") pod \"cinder-operator-controller-manager-8d58dc466-47nf2\" (UID: \"a95e604d-a508-41bc-84e3-78fd32a90ed4\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.214214 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4jl6\" (UniqueName: \"kubernetes.io/projected/c9b3eb51-1ddc-4abc-9d8a-09611ef3983c-kube-api-access-r4jl6\") pod \"barbican-operator-controller-manager-59bc569d95-z9bkx\" (UID: \"c9b3eb51-1ddc-4abc-9d8a-09611ef3983c\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.248805 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.249815 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.260828 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8hf6\" (UniqueName: \"kubernetes.io/projected/f99021be-4dee-4443-938e-ec54f85d61f5-kube-api-access-x8hf6\") pod \"designate-operator-controller-manager-588d4d986b-56xqd\" (UID: \"f99021be-4dee-4443-938e-ec54f85d61f5\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.269164 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-g6xml" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.270737 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.270770 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tvg8\" (UniqueName: \"kubernetes.io/projected/70a65110-b953-43ca-a6ec-5131b4660aea-kube-api-access-5tvg8\") pod \"glance-operator-controller-manager-79df6bcc97-7c7hb\" (UID: \"70a65110-b953-43ca-a6ec-5131b4660aea\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.270915 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.271709 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.273082 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-26k8v" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.278605 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l2z6\" (UniqueName: \"kubernetes.io/projected/a95e604d-a508-41bc-84e3-78fd32a90ed4-kube-api-access-6l2z6\") pod \"cinder-operator-controller-manager-8d58dc466-47nf2\" (UID: \"a95e604d-a508-41bc-84e3-78fd32a90ed4\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.292366 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4jl6\" (UniqueName: \"kubernetes.io/projected/c9b3eb51-1ddc-4abc-9d8a-09611ef3983c-kube-api-access-r4jl6\") pod \"barbican-operator-controller-manager-59bc569d95-z9bkx\" (UID: \"c9b3eb51-1ddc-4abc-9d8a-09611ef3983c\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.292430 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.309016 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.310601 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.320709 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fc4b\" (UniqueName: \"kubernetes.io/projected/ddb36285-c38c-4c75-81ca-4f55f46118e8-kube-api-access-8fc4b\") pod \"horizon-operator-controller-manager-8464cc45fb-75qls\" (UID: \"ddb36285-c38c-4c75-81ca-4f55f46118e8\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.320762 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vngqq\" (UniqueName: \"kubernetes.io/projected/8ed64ffe-e8ec-4e37-a2ba-9a7bfb86e078-kube-api-access-vngqq\") pod \"heat-operator-controller-manager-67dd5f86f5-gxz2f\" (UID: \"8ed64ffe-e8ec-4e37-a2ba-9a7bfb86e078\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.320942 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.349112 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.363078 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.364089 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.370521 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2ntc7" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.383484 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.384284 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.386660 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-6s2nf" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.407047 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.423381 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fc4b\" (UniqueName: \"kubernetes.io/projected/ddb36285-c38c-4c75-81ca-4f55f46118e8-kube-api-access-8fc4b\") pod \"horizon-operator-controller-manager-8464cc45fb-75qls\" (UID: \"ddb36285-c38c-4c75-81ca-4f55f46118e8\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.423430 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vngqq\" (UniqueName: \"kubernetes.io/projected/8ed64ffe-e8ec-4e37-a2ba-9a7bfb86e078-kube-api-access-vngqq\") pod \"heat-operator-controller-manager-67dd5f86f5-gxz2f\" (UID: \"8ed64ffe-e8ec-4e37-a2ba-9a7bfb86e078\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.423463 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4fjf\" (UniqueName: \"kubernetes.io/projected/bf1c744e-50af-44bb-883b-085681d41cff-kube-api-access-l4fjf\") pod \"ironic-operator-controller-manager-6f787dddc9-vcsmh\" (UID: \"bf1c744e-50af-44bb-883b-085681d41cff\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.423501 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5gcz\" (UniqueName: \"kubernetes.io/projected/daa1da6a-6c56-49b4-852d-915636736f69-kube-api-access-z5gcz\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.423532 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.525719 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.526627 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5gcz\" (UniqueName: \"kubernetes.io/projected/daa1da6a-6c56-49b4-852d-915636736f69-kube-api-access-z5gcz\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.526681 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.526769 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z6x7\" (UniqueName: \"kubernetes.io/projected/d933984b-3a5c-4792-a5cb-751cae8865ef-kube-api-access-4z6x7\") pod \"manila-operator-controller-manager-55f864c847-xkcdj\" (UID: \"d933984b-3a5c-4792-a5cb-751cae8865ef\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.526847 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4fjf\" (UniqueName: \"kubernetes.io/projected/bf1c744e-50af-44bb-883b-085681d41cff-kube-api-access-l4fjf\") pod \"ironic-operator-controller-manager-6f787dddc9-vcsmh\" (UID: \"bf1c744e-50af-44bb-883b-085681d41cff\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.526891 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5l47\" (UniqueName: \"kubernetes.io/projected/cf311c3f-f3fd-47b7-9f04-427be533e8d6-kube-api-access-w5l47\") pod \"keystone-operator-controller-manager-768b96df4c-mbb9d\" (UID: \"cf311c3f-f3fd-47b7-9f04-427be533e8d6\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" Mar 20 09:14:18 crc kubenswrapper[4823]: E0320 09:14:18.527342 4823 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:18 crc kubenswrapper[4823]: E0320 09:14:18.527402 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert podName:daa1da6a-6c56-49b4-852d-915636736f69 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:19.02737051 +0000 UTC m=+1025.713266884 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert") pod "infra-operator-controller-manager-577ccd856-ftcjs" (UID: "daa1da6a-6c56-49b4-852d-915636736f69") : secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.536391 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vngqq\" (UniqueName: \"kubernetes.io/projected/8ed64ffe-e8ec-4e37-a2ba-9a7bfb86e078-kube-api-access-vngqq\") pod \"heat-operator-controller-manager-67dd5f86f5-gxz2f\" (UID: \"8ed64ffe-e8ec-4e37-a2ba-9a7bfb86e078\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.547411 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.549823 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.557557 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fc4b\" (UniqueName: \"kubernetes.io/projected/ddb36285-c38c-4c75-81ca-4f55f46118e8-kube-api-access-8fc4b\") pod \"horizon-operator-controller-manager-8464cc45fb-75qls\" (UID: \"ddb36285-c38c-4c75-81ca-4f55f46118e8\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.558064 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.558498 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-tsp8n" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.562282 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4fjf\" (UniqueName: \"kubernetes.io/projected/bf1c744e-50af-44bb-883b-085681d41cff-kube-api-access-l4fjf\") pod \"ironic-operator-controller-manager-6f787dddc9-vcsmh\" (UID: \"bf1c744e-50af-44bb-883b-085681d41cff\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.568380 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.575808 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.576686 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.577902 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.578552 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-fjxb2" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.583089 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.583435 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.583967 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5gcz\" (UniqueName: \"kubernetes.io/projected/daa1da6a-6c56-49b4-852d-915636736f69-kube-api-access-z5gcz\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.587490 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-7qhmc" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.587672 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.606988 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.611786 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.612853 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.616519 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-74t29" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.623083 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.627119 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.631382 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.632872 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-lp942"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.634203 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z6x7\" (UniqueName: \"kubernetes.io/projected/d933984b-3a5c-4792-a5cb-751cae8865ef-kube-api-access-4z6x7\") pod \"manila-operator-controller-manager-55f864c847-xkcdj\" (UID: \"d933984b-3a5c-4792-a5cb-751cae8865ef\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.634359 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ps4q\" (UniqueName: \"kubernetes.io/projected/e19776a8-b3ac-4901-9dbf-ed46ff5b44d9-kube-api-access-8ps4q\") pod \"mariadb-operator-controller-manager-67ccfc9778-bgmxj\" (UID: \"e19776a8-b3ac-4901-9dbf-ed46ff5b44d9\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.634387 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbdg2\" (UniqueName: \"kubernetes.io/projected/570a5a06-26f8-4edb-a891-703fb297adc2-kube-api-access-lbdg2\") pod \"neutron-operator-controller-manager-767865f676-8k6bt\" (UID: \"570a5a06-26f8-4edb-a891-703fb297adc2\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.634453 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8fqm\" (UniqueName: \"kubernetes.io/projected/d8a80347-5759-4eb7-9750-b170fa6e5689-kube-api-access-l8fqm\") pod \"nova-operator-controller-manager-5d488d59fb-28hd6\" (UID: \"d8a80347-5759-4eb7-9750-b170fa6e5689\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.634584 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5l47\" (UniqueName: \"kubernetes.io/projected/cf311c3f-f3fd-47b7-9f04-427be533e8d6-kube-api-access-w5l47\") pod \"keystone-operator-controller-manager-768b96df4c-mbb9d\" (UID: \"cf311c3f-f3fd-47b7-9f04-427be533e8d6\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.634697 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92wnt\" (UniqueName: \"kubernetes.io/projected/638e33c0-1409-4f54-a08b-25c6ef4319a7-kube-api-access-92wnt\") pod \"octavia-operator-controller-manager-5b9f45d989-vft4x\" (UID: \"638e33c0-1409-4f54-a08b-25c6ef4319a7\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.634853 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5plq\" (UniqueName: \"kubernetes.io/projected/ee7cc64f-5303-412e-83b4-322e06279025-kube-api-access-w5plq\") pod \"ovn-operator-controller-manager-884679f54-fdgtg\" (UID: \"ee7cc64f-5303-412e-83b4-322e06279025\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.643652 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-69zpr" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.647787 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-lp942" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.650075 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-lp942"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.652145 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-mttqv" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.666649 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5l47\" (UniqueName: \"kubernetes.io/projected/cf311c3f-f3fd-47b7-9f04-427be533e8d6-kube-api-access-w5l47\") pod \"keystone-operator-controller-manager-768b96df4c-mbb9d\" (UID: \"cf311c3f-f3fd-47b7-9f04-427be533e8d6\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.676031 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.676251 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.676779 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z6x7\" (UniqueName: \"kubernetes.io/projected/d933984b-3a5c-4792-a5cb-751cae8865ef-kube-api-access-4z6x7\") pod \"manila-operator-controller-manager-55f864c847-xkcdj\" (UID: \"d933984b-3a5c-4792-a5cb-751cae8865ef\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.676973 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.684234 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.684428 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-4tmm5" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.692899 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.693766 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.696821 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-l7ncb" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.715308 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.724799 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.728718 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.731907 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.734783 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.735510 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.738180 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbdg2\" (UniqueName: \"kubernetes.io/projected/570a5a06-26f8-4edb-a891-703fb297adc2-kube-api-access-lbdg2\") pod \"neutron-operator-controller-manager-767865f676-8k6bt\" (UID: \"570a5a06-26f8-4edb-a891-703fb297adc2\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.738219 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8fqm\" (UniqueName: \"kubernetes.io/projected/d8a80347-5759-4eb7-9750-b170fa6e5689-kube-api-access-l8fqm\") pod \"nova-operator-controller-manager-5d488d59fb-28hd6\" (UID: \"d8a80347-5759-4eb7-9750-b170fa6e5689\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.738265 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92wnt\" (UniqueName: \"kubernetes.io/projected/638e33c0-1409-4f54-a08b-25c6ef4319a7-kube-api-access-92wnt\") pod \"octavia-operator-controller-manager-5b9f45d989-vft4x\" (UID: \"638e33c0-1409-4f54-a08b-25c6ef4319a7\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.738294 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5plq\" (UniqueName: \"kubernetes.io/projected/ee7cc64f-5303-412e-83b4-322e06279025-kube-api-access-w5plq\") pod \"ovn-operator-controller-manager-884679f54-fdgtg\" (UID: \"ee7cc64f-5303-412e-83b4-322e06279025\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.738338 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ps4q\" (UniqueName: \"kubernetes.io/projected/e19776a8-b3ac-4901-9dbf-ed46ff5b44d9-kube-api-access-8ps4q\") pod \"mariadb-operator-controller-manager-67ccfc9778-bgmxj\" (UID: \"e19776a8-b3ac-4901-9dbf-ed46ff5b44d9\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.740332 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-g4m6m" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.740596 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-mrws4" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.751306 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.761834 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.762314 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbdg2\" (UniqueName: \"kubernetes.io/projected/570a5a06-26f8-4edb-a891-703fb297adc2-kube-api-access-lbdg2\") pod \"neutron-operator-controller-manager-767865f676-8k6bt\" (UID: \"570a5a06-26f8-4edb-a891-703fb297adc2\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.766722 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92wnt\" (UniqueName: \"kubernetes.io/projected/638e33c0-1409-4f54-a08b-25c6ef4319a7-kube-api-access-92wnt\") pod \"octavia-operator-controller-manager-5b9f45d989-vft4x\" (UID: \"638e33c0-1409-4f54-a08b-25c6ef4319a7\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.772567 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.777562 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8fqm\" (UniqueName: \"kubernetes.io/projected/d8a80347-5759-4eb7-9750-b170fa6e5689-kube-api-access-l8fqm\") pod \"nova-operator-controller-manager-5d488d59fb-28hd6\" (UID: \"d8a80347-5759-4eb7-9750-b170fa6e5689\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.782810 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ps4q\" (UniqueName: \"kubernetes.io/projected/e19776a8-b3ac-4901-9dbf-ed46ff5b44d9-kube-api-access-8ps4q\") pod \"mariadb-operator-controller-manager-67ccfc9778-bgmxj\" (UID: \"e19776a8-b3ac-4901-9dbf-ed46ff5b44d9\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.785167 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5plq\" (UniqueName: \"kubernetes.io/projected/ee7cc64f-5303-412e-83b4-322e06279025-kube-api-access-w5plq\") pod \"ovn-operator-controller-manager-884679f54-fdgtg\" (UID: \"ee7cc64f-5303-412e-83b4-322e06279025\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.800122 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.801212 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.801510 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.804613 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-6b28k" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.821248 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.832554 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.840553 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqjh7\" (UniqueName: \"kubernetes.io/projected/091871f2-1d5b-4263-b190-3e6262c4d1b0-kube-api-access-sqjh7\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.858538 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l4kc\" (UniqueName: \"kubernetes.io/projected/98dd6787-8d0c-4612-a18c-0939eaa5a8af-kube-api-access-9l4kc\") pod \"swift-operator-controller-manager-c674c5965-h6dhs\" (UID: \"98dd6787-8d0c-4612-a18c-0939eaa5a8af\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.859066 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5n5l\" (UniqueName: \"kubernetes.io/projected/16e76c27-f7b9-4df7-8cc6-b1b0753613bc-kube-api-access-t5n5l\") pod \"telemetry-operator-controller-manager-d6b694c5-htq9x\" (UID: \"16e76c27-f7b9-4df7-8cc6-b1b0753613bc\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.859216 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.859270 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmkwt\" (UniqueName: \"kubernetes.io/projected/fdc2187a-b568-473f-ab3b-0c1058789b5b-kube-api-access-pmkwt\") pod \"test-operator-controller-manager-5c5cb9c4d7-8bjm7\" (UID: \"fdc2187a-b568-473f-ab3b-0c1058789b5b\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.859325 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z25gv\" (UniqueName: \"kubernetes.io/projected/04565eb9-0d95-440c-a8c0-742a419cae88-kube-api-access-z25gv\") pod \"placement-operator-controller-manager-5784578c99-lp942\" (UID: \"04565eb9-0d95-440c-a8c0-742a419cae88\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-lp942" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.872182 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.895101 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.918532 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.934811 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55958644c4-vk586"] Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.956942 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.959338 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.962031 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.968232 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z25gv\" (UniqueName: \"kubernetes.io/projected/04565eb9-0d95-440c-a8c0-742a419cae88-kube-api-access-z25gv\") pod \"placement-operator-controller-manager-5784578c99-lp942\" (UID: \"04565eb9-0d95-440c-a8c0-742a419cae88\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-lp942" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.968372 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j22d\" (UniqueName: \"kubernetes.io/projected/33c79106-0897-4b9a-ab61-5e873648f9e1-kube-api-access-7j22d\") pod \"watcher-operator-controller-manager-6c4d75f7f9-6z7z9\" (UID: \"33c79106-0897-4b9a-ab61-5e873648f9e1\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.968421 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqjh7\" (UniqueName: \"kubernetes.io/projected/091871f2-1d5b-4263-b190-3e6262c4d1b0-kube-api-access-sqjh7\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.968571 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l4kc\" (UniqueName: \"kubernetes.io/projected/98dd6787-8d0c-4612-a18c-0939eaa5a8af-kube-api-access-9l4kc\") pod \"swift-operator-controller-manager-c674c5965-h6dhs\" (UID: \"98dd6787-8d0c-4612-a18c-0939eaa5a8af\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.968694 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5n5l\" (UniqueName: \"kubernetes.io/projected/16e76c27-f7b9-4df7-8cc6-b1b0753613bc-kube-api-access-t5n5l\") pod \"telemetry-operator-controller-manager-d6b694c5-htq9x\" (UID: \"16e76c27-f7b9-4df7-8cc6-b1b0753613bc\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.968952 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.969047 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmkwt\" (UniqueName: \"kubernetes.io/projected/fdc2187a-b568-473f-ab3b-0c1058789b5b-kube-api-access-pmkwt\") pod \"test-operator-controller-manager-5c5cb9c4d7-8bjm7\" (UID: \"fdc2187a-b568-473f-ab3b-0c1058789b5b\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.970352 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-btv8p" Mar 20 09:14:18 crc kubenswrapper[4823]: E0320 09:14:18.970403 4823 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:18 crc kubenswrapper[4823]: E0320 09:14:18.970481 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert podName:091871f2-1d5b-4263-b190-3e6262c4d1b0 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:19.470449452 +0000 UTC m=+1026.156345826 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" (UID: "091871f2-1d5b-4263-b190-3e6262c4d1b0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.974855 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.975254 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.989184 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" Mar 20 09:14:18 crc kubenswrapper[4823]: I0320 09:14:18.999641 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.007254 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5n5l\" (UniqueName: \"kubernetes.io/projected/16e76c27-f7b9-4df7-8cc6-b1b0753613bc-kube-api-access-t5n5l\") pod \"telemetry-operator-controller-manager-d6b694c5-htq9x\" (UID: \"16e76c27-f7b9-4df7-8cc6-b1b0753613bc\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.007538 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z25gv\" (UniqueName: \"kubernetes.io/projected/04565eb9-0d95-440c-a8c0-742a419cae88-kube-api-access-z25gv\") pod \"placement-operator-controller-manager-5784578c99-lp942\" (UID: \"04565eb9-0d95-440c-a8c0-742a419cae88\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-lp942" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.008954 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmkwt\" (UniqueName: \"kubernetes.io/projected/fdc2187a-b568-473f-ab3b-0c1058789b5b-kube-api-access-pmkwt\") pod \"test-operator-controller-manager-5c5cb9c4d7-8bjm7\" (UID: \"fdc2187a-b568-473f-ab3b-0c1058789b5b\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.014268 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l4kc\" (UniqueName: \"kubernetes.io/projected/98dd6787-8d0c-4612-a18c-0939eaa5a8af-kube-api-access-9l4kc\") pod \"swift-operator-controller-manager-c674c5965-h6dhs\" (UID: \"98dd6787-8d0c-4612-a18c-0939eaa5a8af\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.020580 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55958644c4-vk586"] Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.023421 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-lp942" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.062113 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqjh7\" (UniqueName: \"kubernetes.io/projected/091871f2-1d5b-4263-b190-3e6262c4d1b0-kube-api-access-sqjh7\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.069446 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.070160 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.070303 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.070342 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.070376 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kkth\" (UniqueName: \"kubernetes.io/projected/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-kube-api-access-7kkth\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.070404 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j22d\" (UniqueName: \"kubernetes.io/projected/33c79106-0897-4b9a-ab61-5e873648f9e1-kube-api-access-7j22d\") pod \"watcher-operator-controller-manager-6c4d75f7f9-6z7z9\" (UID: \"33c79106-0897-4b9a-ab61-5e873648f9e1\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.070802 4823 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.070867 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert podName:daa1da6a-6c56-49b4-852d-915636736f69 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:20.070847157 +0000 UTC m=+1026.756743531 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert") pod "infra-operator-controller-manager-577ccd856-ftcjs" (UID: "daa1da6a-6c56-49b4-852d-915636736f69") : secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.087690 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2"] Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.104091 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j22d\" (UniqueName: \"kubernetes.io/projected/33c79106-0897-4b9a-ab61-5e873648f9e1-kube-api-access-7j22d\") pod \"watcher-operator-controller-manager-6c4d75f7f9-6z7z9\" (UID: \"33c79106-0897-4b9a-ab61-5e873648f9e1\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.105084 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.116718 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx"] Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.123015 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.141259 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.171724 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.171855 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kkth\" (UniqueName: \"kubernetes.io/projected/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-kube-api-access-7kkth\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.171938 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.172111 4823 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.172199 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:19.672178151 +0000 UTC m=+1026.358074525 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "metrics-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.172902 4823 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.172982 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:19.672963726 +0000 UTC m=+1026.358860120 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "webhook-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.197177 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kkth\" (UniqueName: \"kubernetes.io/projected/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-kube-api-access-7kkth\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.245367 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb"] Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.261126 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd"] Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.435636 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls"] Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.439015 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2" event={"ID":"a95e604d-a508-41bc-84e3-78fd32a90ed4","Type":"ContainerStarted","Data":"2bd303ab02afe04bd5690ddbd03e382f7d1d38c3f9aaa22a9467309aa79a0a1a"} Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.443521 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd" event={"ID":"f99021be-4dee-4443-938e-ec54f85d61f5","Type":"ContainerStarted","Data":"6f6eabe8d6b529824897be28bf6e97e50312ef21fca83a85233e634348a1dfe9"} Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.448359 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx" event={"ID":"c9b3eb51-1ddc-4abc-9d8a-09611ef3983c","Type":"ContainerStarted","Data":"9ded1681220e91d8933bc39988aaf9be834e8ddca17ffd182a0e06b72cbb5ec7"} Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.450780 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb" event={"ID":"70a65110-b953-43ca-a6ec-5131b4660aea","Type":"ContainerStarted","Data":"62a9315eea2afb01a54dde316e9c83e9243d97baa37446ababbc4a118a035a2a"} Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.465640 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f"] Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.478029 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.478175 4823 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.478228 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert podName:091871f2-1d5b-4263-b190-3e6262c4d1b0 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:20.478214778 +0000 UTC m=+1027.164111142 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" (UID: "091871f2-1d5b-4263-b190-3e6262c4d1b0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.558472 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d"] Mar 20 09:14:19 crc kubenswrapper[4823]: W0320 09:14:19.568636 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf311c3f_f3fd_47b7_9f04_427be533e8d6.slice/crio-d6c470c0c74c58fcb610227929dca2c7fdc91a0a8492f7dd8b2b09e519240576 WatchSource:0}: Error finding container d6c470c0c74c58fcb610227929dca2c7fdc91a0a8492f7dd8b2b09e519240576: Status 404 returned error can't find the container with id d6c470c0c74c58fcb610227929dca2c7fdc91a0a8492f7dd8b2b09e519240576 Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.571048 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh"] Mar 20 09:14:19 crc kubenswrapper[4823]: W0320 09:14:19.578262 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf1c744e_50af_44bb_883b_085681d41cff.slice/crio-ecea28955604758e87a564a34ecb002ee21a4fd3bb859d7bddbb369ade016369 WatchSource:0}: Error finding container ecea28955604758e87a564a34ecb002ee21a4fd3bb859d7bddbb369ade016369: Status 404 returned error can't find the container with id ecea28955604758e87a564a34ecb002ee21a4fd3bb859d7bddbb369ade016369 Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.680543 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.680864 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.680980 4823 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.681026 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:20.6810122 +0000 UTC m=+1027.366908574 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "webhook-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.681324 4823 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.681347 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:20.68134063 +0000 UTC m=+1027.367237004 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "metrics-server-cert" not found Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.714193 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6"] Mar 20 09:14:19 crc kubenswrapper[4823]: W0320 09:14:19.715479 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8a80347_5759_4eb7_9750_b170fa6e5689.slice/crio-134965d6d0ec90bd150e0314b2416fdbcd5bcca67855a5f4909769a50a5f230d WatchSource:0}: Error finding container 134965d6d0ec90bd150e0314b2416fdbcd5bcca67855a5f4909769a50a5f230d: Status 404 returned error can't find the container with id 134965d6d0ec90bd150e0314b2416fdbcd5bcca67855a5f4909769a50a5f230d Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.856413 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj"] Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.866898 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-lp942"] Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.871280 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt"] Mar 20 09:14:19 crc kubenswrapper[4823]: W0320 09:14:19.871316 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode19776a8_b3ac_4901_9dbf_ed46ff5b44d9.slice/crio-3fe53dc7c02c93cdca06f08b6caddaaaa951d31dfde6af9acb86e6a27204e187 WatchSource:0}: Error finding container 3fe53dc7c02c93cdca06f08b6caddaaaa951d31dfde6af9acb86e6a27204e187: Status 404 returned error can't find the container with id 3fe53dc7c02c93cdca06f08b6caddaaaa951d31dfde6af9acb86e6a27204e187 Mar 20 09:14:19 crc kubenswrapper[4823]: W0320 09:14:19.874341 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04565eb9_0d95_440c_a8c0_742a419cae88.slice/crio-749bb2e135c0b99e0492afe918748f84ad90bfb41e0d15d318a0496478effebd WatchSource:0}: Error finding container 749bb2e135c0b99e0492afe918748f84ad90bfb41e0d15d318a0496478effebd: Status 404 returned error can't find the container with id 749bb2e135c0b99e0492afe918748f84ad90bfb41e0d15d318a0496478effebd Mar 20 09:14:19 crc kubenswrapper[4823]: W0320 09:14:19.877626 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod570a5a06_26f8_4edb_a891_703fb297adc2.slice/crio-f6aaf6a874750f4bb58c0e80ca17a7c38a35a2fbe837b15ef95f473998bddaba WatchSource:0}: Error finding container f6aaf6a874750f4bb58c0e80ca17a7c38a35a2fbe837b15ef95f473998bddaba: Status 404 returned error can't find the container with id f6aaf6a874750f4bb58c0e80ca17a7c38a35a2fbe837b15ef95f473998bddaba Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.880637 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs"] Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.885682 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj"] Mar 20 09:14:19 crc kubenswrapper[4823]: W0320 09:14:19.892380 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod638e33c0_1409_4f54_a08b_25c6ef4319a7.slice/crio-87dc9bf3b57ba810873fa61eb140036385ad303aedee26a5fc458bcda2476eaa WatchSource:0}: Error finding container 87dc9bf3b57ba810873fa61eb140036385ad303aedee26a5fc458bcda2476eaa: Status 404 returned error can't find the container with id 87dc9bf3b57ba810873fa61eb140036385ad303aedee26a5fc458bcda2476eaa Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.892770 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4z6x7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-55f864c847-xkcdj_openstack-operators(d933984b-3a5c-4792-a5cb-751cae8865ef): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.893531 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x"] Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.894066 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" podUID="d933984b-3a5c-4792-a5cb-751cae8865ef" Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.895427 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-92wnt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5b9f45d989-vft4x_openstack-operators(638e33c0-1409-4f54-a08b-25c6ef4319a7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 09:14:19 crc kubenswrapper[4823]: E0320 09:14:19.896624 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" podUID="638e33c0-1409-4f54-a08b-25c6ef4319a7" Mar 20 09:14:19 crc kubenswrapper[4823]: I0320 09:14:19.931445 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdc24"] Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.050225 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg"] Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.063044 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9"] Mar 20 09:14:20 crc kubenswrapper[4823]: W0320 09:14:20.063283 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee7cc64f_5303_412e_83b4_322e06279025.slice/crio-138df00b87fcdc05e7fbac3d404fa13c9cae618dc1ae122770477b642d6550bd WatchSource:0}: Error finding container 138df00b87fcdc05e7fbac3d404fa13c9cae618dc1ae122770477b642d6550bd: Status 404 returned error can't find the container with id 138df00b87fcdc05e7fbac3d404fa13c9cae618dc1ae122770477b642d6550bd Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.073400 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7"] Mar 20 09:14:20 crc kubenswrapper[4823]: W0320 09:14:20.076757 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33c79106_0897_4b9a_ab61_5e873648f9e1.slice/crio-639648fca27ea890f5359661ff59bd3f382ffdb3e9c2d36d2f7b649de4a37f07 WatchSource:0}: Error finding container 639648fca27ea890f5359661ff59bd3f382ffdb3e9c2d36d2f7b649de4a37f07: Status 404 returned error can't find the container with id 639648fca27ea890f5359661ff59bd3f382ffdb3e9c2d36d2f7b649de4a37f07 Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.079730 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7j22d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c4d75f7f9-6z7z9_openstack-operators(33c79106-0897-4b9a-ab61-5e873648f9e1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.082202 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" podUID="33c79106-0897-4b9a-ab61-5e873648f9e1" Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.086059 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.086207 4823 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.086259 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert podName:daa1da6a-6c56-49b4-852d-915636736f69 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:22.086243564 +0000 UTC m=+1028.772139938 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert") pod "infra-operator-controller-manager-577ccd856-ftcjs" (UID: "daa1da6a-6c56-49b4-852d-915636736f69") : secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:20 crc kubenswrapper[4823]: W0320 09:14:20.087298 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdc2187a_b568_473f_ab3b_0c1058789b5b.slice/crio-87acd41c5137cf982748364f150b8fe2595cc3880dd4e3cdbd75ca33a58fd22c WatchSource:0}: Error finding container 87acd41c5137cf982748364f150b8fe2595cc3880dd4e3cdbd75ca33a58fd22c: Status 404 returned error can't find the container with id 87acd41c5137cf982748364f150b8fe2595cc3880dd4e3cdbd75ca33a58fd22c Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.087821 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x"] Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.094198 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pmkwt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-8bjm7_openstack-operators(fdc2187a-b568-473f-ab3b-0c1058789b5b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.094610 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t5n5l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-d6b694c5-htq9x_openstack-operators(16e76c27-f7b9-4df7-8cc6-b1b0753613bc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.096569 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" podUID="16e76c27-f7b9-4df7-8cc6-b1b0753613bc" Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.096605 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" podUID="fdc2187a-b568-473f-ab3b-0c1058789b5b" Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.479853 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" event={"ID":"33c79106-0897-4b9a-ab61-5e873648f9e1","Type":"ContainerStarted","Data":"639648fca27ea890f5359661ff59bd3f382ffdb3e9c2d36d2f7b649de4a37f07"} Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.483187 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" podUID="33c79106-0897-4b9a-ab61-5e873648f9e1" Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.483976 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" event={"ID":"d933984b-3a5c-4792-a5cb-751cae8865ef","Type":"ContainerStarted","Data":"1785ca951d1c7cc060851cfaaf37233314cca8bd2c42d2208dfcd16d150d5c46"} Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.485579 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da\\\"\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" podUID="d933984b-3a5c-4792-a5cb-751cae8865ef" Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.485689 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" event={"ID":"16e76c27-f7b9-4df7-8cc6-b1b0753613bc","Type":"ContainerStarted","Data":"a0e5e027d2039edbebb2113fb38e11b9a44f0fc948d189a20f12a6201eb613bf"} Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.487180 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" podUID="16e76c27-f7b9-4df7-8cc6-b1b0753613bc" Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.496521 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.496660 4823 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.496983 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert podName:091871f2-1d5b-4263-b190-3e6262c4d1b0 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:22.496963539 +0000 UTC m=+1029.182859903 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" (UID: "091871f2-1d5b-4263-b190-3e6262c4d1b0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.512849 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs" event={"ID":"98dd6787-8d0c-4612-a18c-0939eaa5a8af","Type":"ContainerStarted","Data":"9b122958365fc51c234874d5402082dbddcc92bfcaafdd815ede7e9941e79559"} Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.518693 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh" event={"ID":"bf1c744e-50af-44bb-883b-085681d41cff","Type":"ContainerStarted","Data":"ecea28955604758e87a564a34ecb002ee21a4fd3bb859d7bddbb369ade016369"} Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.521551 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" event={"ID":"cf311c3f-f3fd-47b7-9f04-427be533e8d6","Type":"ContainerStarted","Data":"d6c470c0c74c58fcb610227929dca2c7fdc91a0a8492f7dd8b2b09e519240576"} Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.523765 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" event={"ID":"ee7cc64f-5303-412e-83b4-322e06279025","Type":"ContainerStarted","Data":"138df00b87fcdc05e7fbac3d404fa13c9cae618dc1ae122770477b642d6550bd"} Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.531833 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj" event={"ID":"e19776a8-b3ac-4901-9dbf-ed46ff5b44d9","Type":"ContainerStarted","Data":"3fe53dc7c02c93cdca06f08b6caddaaaa951d31dfde6af9acb86e6a27204e187"} Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.544557 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" event={"ID":"638e33c0-1409-4f54-a08b-25c6ef4319a7","Type":"ContainerStarted","Data":"87dc9bf3b57ba810873fa61eb140036385ad303aedee26a5fc458bcda2476eaa"} Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.546344 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" podUID="638e33c0-1409-4f54-a08b-25c6ef4319a7" Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.582144 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" event={"ID":"570a5a06-26f8-4edb-a891-703fb297adc2","Type":"ContainerStarted","Data":"f6aaf6a874750f4bb58c0e80ca17a7c38a35a2fbe837b15ef95f473998bddaba"} Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.591622 4823 generic.go:334] "Generic (PLEG): container finished" podID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerID="4d079c14e4fcf643cf3cee590218785fc81c4959da88e6f3a87525aa25abff7a" exitCode=0 Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.591768 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92fnv" event={"ID":"57e15159-c2f8-45a9-b37a-2ed8529d73c7","Type":"ContainerDied","Data":"4d079c14e4fcf643cf3cee590218785fc81c4959da88e6f3a87525aa25abff7a"} Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.595636 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" event={"ID":"fdc2187a-b568-473f-ab3b-0c1058789b5b","Type":"ContainerStarted","Data":"87acd41c5137cf982748364f150b8fe2595cc3880dd4e3cdbd75ca33a58fd22c"} Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.602061 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" podUID="fdc2187a-b568-473f-ab3b-0c1058789b5b" Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.617341 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f" event={"ID":"8ed64ffe-e8ec-4e37-a2ba-9a7bfb86e078","Type":"ContainerStarted","Data":"c1bd01bafdda970db3934e0683d44af952f414eb2d0ea1067e81ebeea8a69e16"} Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.620078 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls" event={"ID":"ddb36285-c38c-4c75-81ca-4f55f46118e8","Type":"ContainerStarted","Data":"e6661291ba821bfd1cea3d654e0ff1f3ada2adf90093b8b3997b18d355301eb2"} Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.623536 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-lp942" event={"ID":"04565eb9-0d95-440c-a8c0-742a419cae88","Type":"ContainerStarted","Data":"749bb2e135c0b99e0492afe918748f84ad90bfb41e0d15d318a0496478effebd"} Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.626449 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mdc24" podUID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" containerName="registry-server" containerID="cri-o://823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902" gracePeriod=2 Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.626782 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" event={"ID":"d8a80347-5759-4eb7-9750-b170fa6e5689","Type":"ContainerStarted","Data":"134965d6d0ec90bd150e0314b2416fdbcd5bcca67855a5f4909769a50a5f230d"} Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.699045 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:20 crc kubenswrapper[4823]: I0320 09:14:20.699135 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.699457 4823 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.699504 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:22.699489973 +0000 UTC m=+1029.385386347 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "webhook-server-cert" not found Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.699530 4823 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:14:20 crc kubenswrapper[4823]: E0320 09:14:20.699614 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:22.699591956 +0000 UTC m=+1029.385488410 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "metrics-server-cert" not found Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.170733 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.311361 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-utilities\") pod \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.311473 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-catalog-content\") pod \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.311583 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2kcb\" (UniqueName: \"kubernetes.io/projected/d1ce5c2e-da8c-4004-ab2a-790f563accfb-kube-api-access-b2kcb\") pod \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\" (UID: \"d1ce5c2e-da8c-4004-ab2a-790f563accfb\") " Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.314771 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-utilities" (OuterVolumeSpecName: "utilities") pod "d1ce5c2e-da8c-4004-ab2a-790f563accfb" (UID: "d1ce5c2e-da8c-4004-ab2a-790f563accfb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.319548 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1ce5c2e-da8c-4004-ab2a-790f563accfb-kube-api-access-b2kcb" (OuterVolumeSpecName: "kube-api-access-b2kcb") pod "d1ce5c2e-da8c-4004-ab2a-790f563accfb" (UID: "d1ce5c2e-da8c-4004-ab2a-790f563accfb"). InnerVolumeSpecName "kube-api-access-b2kcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.348228 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1ce5c2e-da8c-4004-ab2a-790f563accfb" (UID: "d1ce5c2e-da8c-4004-ab2a-790f563accfb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.413693 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2kcb\" (UniqueName: \"kubernetes.io/projected/d1ce5c2e-da8c-4004-ab2a-790f563accfb-kube-api-access-b2kcb\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.413751 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.413765 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ce5c2e-da8c-4004-ab2a-790f563accfb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.648902 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92fnv" event={"ID":"57e15159-c2f8-45a9-b37a-2ed8529d73c7","Type":"ContainerStarted","Data":"46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7"} Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.677853 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-92fnv" podStartSLOduration=2.90796717 podStartE2EDuration="6.677837956s" podCreationTimestamp="2026-03-20 09:14:15 +0000 UTC" firstStartedPulling="2026-03-20 09:14:17.399953057 +0000 UTC m=+1024.085849421" lastFinishedPulling="2026-03-20 09:14:21.169823833 +0000 UTC m=+1027.855720207" observedRunningTime="2026-03-20 09:14:21.677600988 +0000 UTC m=+1028.363497362" watchObservedRunningTime="2026-03-20 09:14:21.677837956 +0000 UTC m=+1028.363734330" Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.679930 4823 generic.go:334] "Generic (PLEG): container finished" podID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" containerID="823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902" exitCode=0 Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.681261 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdc24" Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.682084 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdc24" event={"ID":"d1ce5c2e-da8c-4004-ab2a-790f563accfb","Type":"ContainerDied","Data":"823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902"} Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.682112 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdc24" event={"ID":"d1ce5c2e-da8c-4004-ab2a-790f563accfb","Type":"ContainerDied","Data":"122f00e5a9f4537a184457b34f3b4f57b751aee4fd46440de06a34ed98e338f4"} Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.682129 4823 scope.go:117] "RemoveContainer" containerID="823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902" Mar 20 09:14:21 crc kubenswrapper[4823]: E0320 09:14:21.687194 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da\\\"\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" podUID="d933984b-3a5c-4792-a5cb-751cae8865ef" Mar 20 09:14:21 crc kubenswrapper[4823]: E0320 09:14:21.687941 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" podUID="33c79106-0897-4b9a-ab61-5e873648f9e1" Mar 20 09:14:21 crc kubenswrapper[4823]: E0320 09:14:21.688064 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" podUID="16e76c27-f7b9-4df7-8cc6-b1b0753613bc" Mar 20 09:14:21 crc kubenswrapper[4823]: E0320 09:14:21.688905 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" podUID="fdc2187a-b568-473f-ab3b-0c1058789b5b" Mar 20 09:14:21 crc kubenswrapper[4823]: E0320 09:14:21.698995 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" podUID="638e33c0-1409-4f54-a08b-25c6ef4319a7" Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.790293 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdc24"] Mar 20 09:14:21 crc kubenswrapper[4823]: I0320 09:14:21.794711 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdc24"] Mar 20 09:14:22 crc kubenswrapper[4823]: I0320 09:14:22.109727 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" path="/var/lib/kubelet/pods/d1ce5c2e-da8c-4004-ab2a-790f563accfb/volumes" Mar 20 09:14:22 crc kubenswrapper[4823]: I0320 09:14:22.129082 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:22 crc kubenswrapper[4823]: E0320 09:14:22.129317 4823 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:22 crc kubenswrapper[4823]: E0320 09:14:22.129386 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert podName:daa1da6a-6c56-49b4-852d-915636736f69 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:26.129369141 +0000 UTC m=+1032.815265515 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert") pod "infra-operator-controller-manager-577ccd856-ftcjs" (UID: "daa1da6a-6c56-49b4-852d-915636736f69") : secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:22 crc kubenswrapper[4823]: I0320 09:14:22.535465 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:22 crc kubenswrapper[4823]: E0320 09:14:22.535657 4823 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:22 crc kubenswrapper[4823]: E0320 09:14:22.535969 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert podName:091871f2-1d5b-4263-b190-3e6262c4d1b0 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:26.535950197 +0000 UTC m=+1033.221846571 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" (UID: "091871f2-1d5b-4263-b190-3e6262c4d1b0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:22 crc kubenswrapper[4823]: I0320 09:14:22.740470 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:22 crc kubenswrapper[4823]: I0320 09:14:22.740548 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:22 crc kubenswrapper[4823]: E0320 09:14:22.740657 4823 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:14:22 crc kubenswrapper[4823]: E0320 09:14:22.740710 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:26.740692599 +0000 UTC m=+1033.426588993 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "webhook-server-cert" not found Mar 20 09:14:22 crc kubenswrapper[4823]: E0320 09:14:22.741047 4823 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:14:22 crc kubenswrapper[4823]: E0320 09:14:22.741074 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:26.741066801 +0000 UTC m=+1033.426963185 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "metrics-server-cert" not found Mar 20 09:14:25 crc kubenswrapper[4823]: I0320 09:14:25.882478 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:25 crc kubenswrapper[4823]: I0320 09:14:25.882634 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:25 crc kubenswrapper[4823]: I0320 09:14:25.927243 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:26 crc kubenswrapper[4823]: I0320 09:14:26.205807 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:26 crc kubenswrapper[4823]: E0320 09:14:26.205939 4823 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:26 crc kubenswrapper[4823]: E0320 09:14:26.206105 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert podName:daa1da6a-6c56-49b4-852d-915636736f69 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:34.206085169 +0000 UTC m=+1040.891981553 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert") pod "infra-operator-controller-manager-577ccd856-ftcjs" (UID: "daa1da6a-6c56-49b4-852d-915636736f69") : secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:26 crc kubenswrapper[4823]: I0320 09:14:26.612136 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:26 crc kubenswrapper[4823]: E0320 09:14:26.612283 4823 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:26 crc kubenswrapper[4823]: E0320 09:14:26.612339 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert podName:091871f2-1d5b-4263-b190-3e6262c4d1b0 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:34.612325114 +0000 UTC m=+1041.298221488 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" (UID: "091871f2-1d5b-4263-b190-3e6262c4d1b0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:26 crc kubenswrapper[4823]: I0320 09:14:26.761801 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:26 crc kubenswrapper[4823]: I0320 09:14:26.814838 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:26 crc kubenswrapper[4823]: I0320 09:14:26.814971 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:26 crc kubenswrapper[4823]: E0320 09:14:26.815030 4823 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:14:26 crc kubenswrapper[4823]: E0320 09:14:26.815103 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:34.815083085 +0000 UTC m=+1041.500979459 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "webhook-server-cert" not found Mar 20 09:14:26 crc kubenswrapper[4823]: E0320 09:14:26.815114 4823 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:14:26 crc kubenswrapper[4823]: E0320 09:14:26.815171 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:34.815154017 +0000 UTC m=+1041.501050451 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "metrics-server-cert" not found Mar 20 09:14:27 crc kubenswrapper[4823]: I0320 09:14:27.331497 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-92fnv"] Mar 20 09:14:28 crc kubenswrapper[4823]: I0320 09:14:28.734867 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-92fnv" podUID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerName="registry-server" containerID="cri-o://46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7" gracePeriod=2 Mar 20 09:14:29 crc kubenswrapper[4823]: I0320 09:14:29.750135 4823 generic.go:334] "Generic (PLEG): container finished" podID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerID="46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7" exitCode=0 Mar 20 09:14:29 crc kubenswrapper[4823]: I0320 09:14:29.750186 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92fnv" event={"ID":"57e15159-c2f8-45a9-b37a-2ed8529d73c7","Type":"ContainerDied","Data":"46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7"} Mar 20 09:14:31 crc kubenswrapper[4823]: I0320 09:14:31.682819 4823 scope.go:117] "RemoveContainer" containerID="1370ae68afa59b2b287edd65c55521ba880b60ace22809854da5ada82aaee09c" Mar 20 09:14:32 crc kubenswrapper[4823]: E0320 09:14:32.583461 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56" Mar 20 09:14:32 crc kubenswrapper[4823]: E0320 09:14:32.583716 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w5l47,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-768b96df4c-mbb9d_openstack-operators(cf311c3f-f3fd-47b7-9f04-427be533e8d6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:14:32 crc kubenswrapper[4823]: E0320 09:14:32.585368 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" podUID="cf311c3f-f3fd-47b7-9f04-427be533e8d6" Mar 20 09:14:32 crc kubenswrapper[4823]: E0320 09:14:32.768621 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" podUID="cf311c3f-f3fd-47b7-9f04-427be533e8d6" Mar 20 09:14:33 crc kubenswrapper[4823]: E0320 09:14:33.169299 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55" Mar 20 09:14:33 crc kubenswrapper[4823]: E0320 09:14:33.169529 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w5plq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-884679f54-fdgtg_openstack-operators(ee7cc64f-5303-412e-83b4-322e06279025): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:14:33 crc kubenswrapper[4823]: E0320 09:14:33.170768 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" podUID="ee7cc64f-5303-412e-83b4-322e06279025" Mar 20 09:14:33 crc kubenswrapper[4823]: E0320 09:14:33.593231 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a" Mar 20 09:14:33 crc kubenswrapper[4823]: E0320 09:14:33.593429 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lbdg2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-767865f676-8k6bt_openstack-operators(570a5a06-26f8-4edb-a891-703fb297adc2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:14:33 crc kubenswrapper[4823]: E0320 09:14:33.594887 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" podUID="570a5a06-26f8-4edb-a891-703fb297adc2" Mar 20 09:14:33 crc kubenswrapper[4823]: I0320 09:14:33.598198 4823 scope.go:117] "RemoveContainer" containerID="933dc4e57a873cda6b684cdc8d732e377a16cf8c796fe00e1807c57227303c1f" Mar 20 09:14:33 crc kubenswrapper[4823]: E0320 09:14:33.782594 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" podUID="ee7cc64f-5303-412e-83b4-322e06279025" Mar 20 09:14:33 crc kubenswrapper[4823]: E0320 09:14:33.783324 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" podUID="570a5a06-26f8-4edb-a891-703fb297adc2" Mar 20 09:14:34 crc kubenswrapper[4823]: I0320 09:14:34.270458 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:34 crc kubenswrapper[4823]: E0320 09:14:34.270612 4823 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:34 crc kubenswrapper[4823]: E0320 09:14:34.270667 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert podName:daa1da6a-6c56-49b4-852d-915636736f69 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:50.270653326 +0000 UTC m=+1056.956549700 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert") pod "infra-operator-controller-manager-577ccd856-ftcjs" (UID: "daa1da6a-6c56-49b4-852d-915636736f69") : secret "infra-operator-webhook-server-cert" not found Mar 20 09:14:34 crc kubenswrapper[4823]: I0320 09:14:34.677286 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:34 crc kubenswrapper[4823]: E0320 09:14:34.677521 4823 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:34 crc kubenswrapper[4823]: E0320 09:14:34.677581 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert podName:091871f2-1d5b-4263-b190-3e6262c4d1b0 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:50.677564765 +0000 UTC m=+1057.363461149 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" (UID: "091871f2-1d5b-4263-b190-3e6262c4d1b0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:14:34 crc kubenswrapper[4823]: I0320 09:14:34.880259 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:34 crc kubenswrapper[4823]: I0320 09:14:34.880401 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:34 crc kubenswrapper[4823]: E0320 09:14:34.880533 4823 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:14:34 crc kubenswrapper[4823]: E0320 09:14:34.880580 4823 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:14:34 crc kubenswrapper[4823]: E0320 09:14:34.880623 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:50.880602242 +0000 UTC m=+1057.566498626 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "metrics-server-cert" not found Mar 20 09:14:34 crc kubenswrapper[4823]: E0320 09:14:34.880649 4823 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs podName:dfc33271-a911-4017-bb46-3d6f5ce6ddf1 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:50.880631153 +0000 UTC m=+1057.566527617 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-vk586" (UID: "dfc33271-a911-4017-bb46-3d6f5ce6ddf1") : secret "webhook-server-cert" not found Mar 20 09:14:35 crc kubenswrapper[4823]: E0320 09:14:35.883303 4823 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7 is running failed: container process not found" containerID="46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 09:14:35 crc kubenswrapper[4823]: E0320 09:14:35.883690 4823 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7 is running failed: container process not found" containerID="46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 09:14:35 crc kubenswrapper[4823]: E0320 09:14:35.884076 4823 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7 is running failed: container process not found" containerID="46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 09:14:35 crc kubenswrapper[4823]: E0320 09:14:35.884117 4823 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-92fnv" podUID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerName="registry-server" Mar 20 09:14:38 crc kubenswrapper[4823]: E0320 09:14:38.034271 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a" Mar 20 09:14:38 crc kubenswrapper[4823]: E0320 09:14:38.034814 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l8fqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5d488d59fb-28hd6_openstack-operators(d8a80347-5759-4eb7-9750-b170fa6e5689): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:14:38 crc kubenswrapper[4823]: E0320 09:14:38.036074 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" podUID="d8a80347-5759-4eb7-9750-b170fa6e5689" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.577066 4823 scope.go:117] "RemoveContainer" containerID="47d0af9108752feb6de2614ae3ef6b312f4709c87efc10b180c1a0b2bdc3f167" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.585759 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.634259 4823 scope.go:117] "RemoveContainer" containerID="823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902" Mar 20 09:14:38 crc kubenswrapper[4823]: E0320 09:14:38.637917 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902\": container with ID starting with 823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902 not found: ID does not exist" containerID="823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.637961 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902"} err="failed to get container status \"823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902\": rpc error: code = NotFound desc = could not find container \"823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902\": container with ID starting with 823e4ee25b20a50b53421dae073180d618e507fb83f3e6d7800d443119999902 not found: ID does not exist" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.637987 4823 scope.go:117] "RemoveContainer" containerID="933dc4e57a873cda6b684cdc8d732e377a16cf8c796fe00e1807c57227303c1f" Mar 20 09:14:38 crc kubenswrapper[4823]: E0320 09:14:38.638473 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"933dc4e57a873cda6b684cdc8d732e377a16cf8c796fe00e1807c57227303c1f\": container with ID starting with 933dc4e57a873cda6b684cdc8d732e377a16cf8c796fe00e1807c57227303c1f not found: ID does not exist" containerID="933dc4e57a873cda6b684cdc8d732e377a16cf8c796fe00e1807c57227303c1f" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.638526 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"933dc4e57a873cda6b684cdc8d732e377a16cf8c796fe00e1807c57227303c1f"} err="failed to get container status \"933dc4e57a873cda6b684cdc8d732e377a16cf8c796fe00e1807c57227303c1f\": rpc error: code = NotFound desc = could not find container \"933dc4e57a873cda6b684cdc8d732e377a16cf8c796fe00e1807c57227303c1f\": container with ID starting with 933dc4e57a873cda6b684cdc8d732e377a16cf8c796fe00e1807c57227303c1f not found: ID does not exist" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.638557 4823 scope.go:117] "RemoveContainer" containerID="47d0af9108752feb6de2614ae3ef6b312f4709c87efc10b180c1a0b2bdc3f167" Mar 20 09:14:38 crc kubenswrapper[4823]: E0320 09:14:38.639793 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47d0af9108752feb6de2614ae3ef6b312f4709c87efc10b180c1a0b2bdc3f167\": container with ID starting with 47d0af9108752feb6de2614ae3ef6b312f4709c87efc10b180c1a0b2bdc3f167 not found: ID does not exist" containerID="47d0af9108752feb6de2614ae3ef6b312f4709c87efc10b180c1a0b2bdc3f167" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.639819 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47d0af9108752feb6de2614ae3ef6b312f4709c87efc10b180c1a0b2bdc3f167"} err="failed to get container status \"47d0af9108752feb6de2614ae3ef6b312f4709c87efc10b180c1a0b2bdc3f167\": rpc error: code = NotFound desc = could not find container \"47d0af9108752feb6de2614ae3ef6b312f4709c87efc10b180c1a0b2bdc3f167\": container with ID starting with 47d0af9108752feb6de2614ae3ef6b312f4709c87efc10b180c1a0b2bdc3f167 not found: ID does not exist" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.738505 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-catalog-content\") pod \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.738648 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-utilities\") pod \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.738677 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4j9gg\" (UniqueName: \"kubernetes.io/projected/57e15159-c2f8-45a9-b37a-2ed8529d73c7-kube-api-access-4j9gg\") pod \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\" (UID: \"57e15159-c2f8-45a9-b37a-2ed8529d73c7\") " Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.739647 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-utilities" (OuterVolumeSpecName: "utilities") pod "57e15159-c2f8-45a9-b37a-2ed8529d73c7" (UID: "57e15159-c2f8-45a9-b37a-2ed8529d73c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.748965 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e15159-c2f8-45a9-b37a-2ed8529d73c7-kube-api-access-4j9gg" (OuterVolumeSpecName: "kube-api-access-4j9gg") pod "57e15159-c2f8-45a9-b37a-2ed8529d73c7" (UID: "57e15159-c2f8-45a9-b37a-2ed8529d73c7"). InnerVolumeSpecName "kube-api-access-4j9gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.808302 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57e15159-c2f8-45a9-b37a-2ed8529d73c7" (UID: "57e15159-c2f8-45a9-b37a-2ed8529d73c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.825521 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2" event={"ID":"a95e604d-a508-41bc-84e3-78fd32a90ed4","Type":"ContainerStarted","Data":"be0d74a0b3e3ec891a7651376991a486468ca19604377b1a85fd6dc896f57f91"} Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.826006 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.840528 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4j9gg\" (UniqueName: \"kubernetes.io/projected/57e15159-c2f8-45a9-b37a-2ed8529d73c7-kube-api-access-4j9gg\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.840564 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.840578 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57e15159-c2f8-45a9-b37a-2ed8529d73c7-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.845481 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2" podStartSLOduration=2.472890687 podStartE2EDuration="21.845464395s" podCreationTimestamp="2026-03-20 09:14:17 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.152287552 +0000 UTC m=+1025.838183926" lastFinishedPulling="2026-03-20 09:14:38.52486126 +0000 UTC m=+1045.210757634" observedRunningTime="2026-03-20 09:14:38.839728638 +0000 UTC m=+1045.525625022" watchObservedRunningTime="2026-03-20 09:14:38.845464395 +0000 UTC m=+1045.531360769" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.855882 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92fnv" event={"ID":"57e15159-c2f8-45a9-b37a-2ed8529d73c7","Type":"ContainerDied","Data":"c0eb1d25654d9de886a0cfea9f32bcb123bc86c80eb0273586dee5d06574ba36"} Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.855932 4823 scope.go:117] "RemoveContainer" containerID="46234170c0c845d1a247b34da4b8a3225c318018b515712e071800ed108679a7" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.856007 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92fnv" Mar 20 09:14:38 crc kubenswrapper[4823]: E0320 09:14:38.866572 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" podUID="d8a80347-5759-4eb7-9750-b170fa6e5689" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.905319 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-92fnv"] Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.912806 4823 scope.go:117] "RemoveContainer" containerID="4d079c14e4fcf643cf3cee590218785fc81c4959da88e6f3a87525aa25abff7a" Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.926083 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-92fnv"] Mar 20 09:14:38 crc kubenswrapper[4823]: I0320 09:14:38.961102 4823 scope.go:117] "RemoveContainer" containerID="3a1793ae11f311dc48be360d6e1b3fda6998de43b9b18d263073ae39f8d7f6b5" Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.879646 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj" event={"ID":"e19776a8-b3ac-4901-9dbf-ed46ff5b44d9","Type":"ContainerStarted","Data":"23234f8bee48a3b89708cb7cceffe0dc6da6fb3c2baa6ae1ecd6ea0aff3f62c4"} Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.880707 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj" Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.887281 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" event={"ID":"fdc2187a-b568-473f-ab3b-0c1058789b5b","Type":"ContainerStarted","Data":"e250c41d7bbcafb94de4ec965fcc5049e5b9fe42d0811c4b3edd7c62841faaa4"} Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.887757 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.893011 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-lp942" event={"ID":"04565eb9-0d95-440c-a8c0-742a419cae88","Type":"ContainerStarted","Data":"ae80345f6a386c42ef237ac014730f27014d5da1ac0bbc1f0c85a4b85685e483"} Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.893548 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5784578c99-lp942" Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.898417 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb" event={"ID":"70a65110-b953-43ca-a6ec-5131b4660aea","Type":"ContainerStarted","Data":"07a2170b6e95d22df4165375e170a1c145a70b035d8add6702100bf0597c272e"} Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.899119 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb" Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.902994 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj" podStartSLOduration=3.247259461 podStartE2EDuration="21.902976655s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.873849252 +0000 UTC m=+1026.559745626" lastFinishedPulling="2026-03-20 09:14:38.529566446 +0000 UTC m=+1045.215462820" observedRunningTime="2026-03-20 09:14:39.898240409 +0000 UTC m=+1046.584136783" watchObservedRunningTime="2026-03-20 09:14:39.902976655 +0000 UTC m=+1046.588873029" Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.923364 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd" event={"ID":"f99021be-4dee-4443-938e-ec54f85d61f5","Type":"ContainerStarted","Data":"0aa20e6888d19e91775294103d1bd321da8a69e8d6cd9fad7bb17c62c6d6fc9c"} Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.923978 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd" Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.930107 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f" event={"ID":"8ed64ffe-e8ec-4e37-a2ba-9a7bfb86e078","Type":"ContainerStarted","Data":"9d49866a247bfa793d1cd680c3de65e10fce96ab702f339cbbc2ae3c7634ccc8"} Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.930684 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f" Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.937584 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5784578c99-lp942" podStartSLOduration=3.283294215 podStartE2EDuration="21.937570873s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.876866486 +0000 UTC m=+1026.562762860" lastFinishedPulling="2026-03-20 09:14:38.531143144 +0000 UTC m=+1045.217039518" observedRunningTime="2026-03-20 09:14:39.930968789 +0000 UTC m=+1046.616865163" watchObservedRunningTime="2026-03-20 09:14:39.937570873 +0000 UTC m=+1046.623467247" Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.953023 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls" event={"ID":"ddb36285-c38c-4c75-81ca-4f55f46118e8","Type":"ContainerStarted","Data":"4b4099415cbc2cbe2fe3d14700c6b37e0d5a6a350eb71d9d77570f0a8102809d"} Mar 20 09:14:39 crc kubenswrapper[4823]: I0320 09:14:39.953640 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls" Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.000797 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" podStartSLOduration=3.475274056 podStartE2EDuration="22.000777203s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:20.090514777 +0000 UTC m=+1026.776411151" lastFinishedPulling="2026-03-20 09:14:38.616017924 +0000 UTC m=+1045.301914298" observedRunningTime="2026-03-20 09:14:39.964129452 +0000 UTC m=+1046.650025826" watchObservedRunningTime="2026-03-20 09:14:40.000777203 +0000 UTC m=+1046.686673587" Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.010990 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx" Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.011036 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx" event={"ID":"c9b3eb51-1ddc-4abc-9d8a-09611ef3983c","Type":"ContainerStarted","Data":"c4026f324475c077ffb0b3634eca6085c666682b03f33f92e42c78147771f5ff"} Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.028139 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs" event={"ID":"98dd6787-8d0c-4612-a18c-0939eaa5a8af","Type":"ContainerStarted","Data":"33c521a683fafa81c5d1696c07bff5e544bed18105859f68241a05c5b255f388"} Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.028540 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs" Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.048701 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb" podStartSLOduration=2.830027527 podStartE2EDuration="22.048679152s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.312494239 +0000 UTC m=+1025.998390613" lastFinishedPulling="2026-03-20 09:14:38.531145864 +0000 UTC m=+1045.217042238" observedRunningTime="2026-03-20 09:14:40.022479713 +0000 UTC m=+1046.708376087" watchObservedRunningTime="2026-03-20 09:14:40.048679152 +0000 UTC m=+1046.734575526" Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.074063 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh" event={"ID":"bf1c744e-50af-44bb-883b-085681d41cff","Type":"ContainerStarted","Data":"45cce08d0fd13fb33ca0ab7c03a25fc58481bb76d26b1897f0991051c70bc3cb"} Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.074241 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh" Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.088685 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls" podStartSLOduration=3.02191048 podStartE2EDuration="22.088665186s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.464694147 +0000 UTC m=+1026.150590521" lastFinishedPulling="2026-03-20 09:14:38.531448853 +0000 UTC m=+1045.217345227" observedRunningTime="2026-03-20 09:14:40.074024724 +0000 UTC m=+1046.759921098" watchObservedRunningTime="2026-03-20 09:14:40.088665186 +0000 UTC m=+1046.774561560" Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.115255 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f" podStartSLOduration=3.0738923 podStartE2EDuration="22.115234516s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.486057322 +0000 UTC m=+1026.171953696" lastFinishedPulling="2026-03-20 09:14:38.527399538 +0000 UTC m=+1045.213295912" observedRunningTime="2026-03-20 09:14:40.110854641 +0000 UTC m=+1046.796751005" watchObservedRunningTime="2026-03-20 09:14:40.115234516 +0000 UTC m=+1046.801130890" Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.115646 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" path="/var/lib/kubelet/pods/57e15159-c2f8-45a9-b37a-2ed8529d73c7/volumes" Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.178803 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd" podStartSLOduration=2.961910538 podStartE2EDuration="22.178773658s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.312111907 +0000 UTC m=+1025.998008281" lastFinishedPulling="2026-03-20 09:14:38.528975027 +0000 UTC m=+1045.214871401" observedRunningTime="2026-03-20 09:14:40.144579472 +0000 UTC m=+1046.830475846" watchObservedRunningTime="2026-03-20 09:14:40.178773658 +0000 UTC m=+1046.864670032" Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.226566 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs" podStartSLOduration=3.5810855090000002 podStartE2EDuration="22.226548962s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.885244367 +0000 UTC m=+1026.571140741" lastFinishedPulling="2026-03-20 09:14:38.53070782 +0000 UTC m=+1045.216604194" observedRunningTime="2026-03-20 09:14:40.22584409 +0000 UTC m=+1046.911740464" watchObservedRunningTime="2026-03-20 09:14:40.226548962 +0000 UTC m=+1046.912445336" Mar 20 09:14:40 crc kubenswrapper[4823]: I0320 09:14:40.230625 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx" podStartSLOduration=3.8587942330000002 podStartE2EDuration="23.230614707s" podCreationTimestamp="2026-03-20 09:14:17 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.152884581 +0000 UTC m=+1025.838780955" lastFinishedPulling="2026-03-20 09:14:38.524705065 +0000 UTC m=+1045.210601429" observedRunningTime="2026-03-20 09:14:40.177292911 +0000 UTC m=+1046.863189295" watchObservedRunningTime="2026-03-20 09:14:40.230614707 +0000 UTC m=+1046.916511081" Mar 20 09:14:48 crc kubenswrapper[4823]: I0320 09:14:48.114488 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh" podStartSLOduration=11.173493636 podStartE2EDuration="30.114467408s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.587320904 +0000 UTC m=+1026.273217278" lastFinishedPulling="2026-03-20 09:14:38.528294676 +0000 UTC m=+1045.214191050" observedRunningTime="2026-03-20 09:14:40.270710095 +0000 UTC m=+1046.956606469" watchObservedRunningTime="2026-03-20 09:14:48.114467408 +0000 UTC m=+1054.800363792" Mar 20 09:14:48 crc kubenswrapper[4823]: I0320 09:14:48.314442 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-z9bkx" Mar 20 09:14:48 crc kubenswrapper[4823]: I0320 09:14:48.326360 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-47nf2" Mar 20 09:14:48 crc kubenswrapper[4823]: I0320 09:14:48.352983 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-56xqd" Mar 20 09:14:48 crc kubenswrapper[4823]: I0320 09:14:48.409605 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-7c7hb" Mar 20 09:14:48 crc kubenswrapper[4823]: I0320 09:14:48.775426 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-gxz2f" Mar 20 09:14:48 crc kubenswrapper[4823]: I0320 09:14:48.831821 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-75qls" Mar 20 09:14:48 crc kubenswrapper[4823]: I0320 09:14:48.833589 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-vcsmh" Mar 20 09:14:48 crc kubenswrapper[4823]: I0320 09:14:48.963590 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-bgmxj" Mar 20 09:14:49 crc kubenswrapper[4823]: I0320 09:14:49.026486 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5784578c99-lp942" Mar 20 09:14:49 crc kubenswrapper[4823]: I0320 09:14:49.072074 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-c674c5965-h6dhs" Mar 20 09:14:49 crc kubenswrapper[4823]: I0320 09:14:49.125940 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8bjm7" Mar 20 09:14:50 crc kubenswrapper[4823]: I0320 09:14:50.357796 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:50 crc kubenswrapper[4823]: I0320 09:14:50.363415 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/daa1da6a-6c56-49b4-852d-915636736f69-cert\") pod \"infra-operator-controller-manager-577ccd856-ftcjs\" (UID: \"daa1da6a-6c56-49b4-852d-915636736f69\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:50 crc kubenswrapper[4823]: I0320 09:14:50.471902 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:50 crc kubenswrapper[4823]: I0320 09:14:50.762939 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:50 crc kubenswrapper[4823]: I0320 09:14:50.770075 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/091871f2-1d5b-4263-b190-3e6262c4d1b0-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-rtmw4\" (UID: \"091871f2-1d5b-4263-b190-3e6262c4d1b0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:50 crc kubenswrapper[4823]: E0320 09:14:50.827762 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a" Mar 20 09:14:50 crc kubenswrapper[4823]: E0320 09:14:50.828307 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-92wnt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5b9f45d989-vft4x_openstack-operators(638e33c0-1409-4f54-a08b-25c6ef4319a7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:14:50 crc kubenswrapper[4823]: E0320 09:14:50.830119 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" podUID="638e33c0-1409-4f54-a08b-25c6ef4319a7" Mar 20 09:14:50 crc kubenswrapper[4823]: I0320 09:14:50.853033 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:50 crc kubenswrapper[4823]: I0320 09:14:50.964820 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:50 crc kubenswrapper[4823]: I0320 09:14:50.964934 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:50 crc kubenswrapper[4823]: I0320 09:14:50.970681 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:50 crc kubenswrapper[4823]: I0320 09:14:50.971361 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dfc33271-a911-4017-bb46-3d6f5ce6ddf1-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-vk586\" (UID: \"dfc33271-a911-4017-bb46-3d6f5ce6ddf1\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:51 crc kubenswrapper[4823]: I0320 09:14:51.178103 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:51 crc kubenswrapper[4823]: I0320 09:14:51.195227 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" event={"ID":"16e76c27-f7b9-4df7-8cc6-b1b0753613bc","Type":"ContainerStarted","Data":"a9548c298014a2f20ad9965083a878bac4261f0edfc959df0f1cd2ab80ac3b5e"} Mar 20 09:14:51 crc kubenswrapper[4823]: I0320 09:14:51.195417 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" Mar 20 09:14:51 crc kubenswrapper[4823]: I0320 09:14:51.197586 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" event={"ID":"cf311c3f-f3fd-47b7-9f04-427be533e8d6","Type":"ContainerStarted","Data":"8faf9dcaa2d5c0f37ac8b3388be8ba1d6eae2fd5b8ffab792e0f26bf92c89331"} Mar 20 09:14:51 crc kubenswrapper[4823]: I0320 09:14:51.198017 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" Mar 20 09:14:51 crc kubenswrapper[4823]: I0320 09:14:51.207943 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" podStartSLOduration=2.360013486 podStartE2EDuration="33.207929096s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:20.09446183 +0000 UTC m=+1026.780358204" lastFinishedPulling="2026-03-20 09:14:50.94237743 +0000 UTC m=+1057.628273814" observedRunningTime="2026-03-20 09:14:51.206576094 +0000 UTC m=+1057.892472468" watchObservedRunningTime="2026-03-20 09:14:51.207929096 +0000 UTC m=+1057.893825470" Mar 20 09:14:51 crc kubenswrapper[4823]: I0320 09:14:51.222168 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" podStartSLOduration=1.855086966 podStartE2EDuration="33.222152395s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.579883462 +0000 UTC m=+1026.265779836" lastFinishedPulling="2026-03-20 09:14:50.946948901 +0000 UTC m=+1057.632845265" observedRunningTime="2026-03-20 09:14:51.221506345 +0000 UTC m=+1057.907402709" watchObservedRunningTime="2026-03-20 09:14:51.222152395 +0000 UTC m=+1057.908048769" Mar 20 09:14:51 crc kubenswrapper[4823]: I0320 09:14:51.426332 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs"] Mar 20 09:14:51 crc kubenswrapper[4823]: W0320 09:14:51.435779 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddaa1da6a_6c56_49b4_852d_915636736f69.slice/crio-7e2b851f5b04600977b2f01f3bdc7559d5afe075a51fc053e5b025534f710348 WatchSource:0}: Error finding container 7e2b851f5b04600977b2f01f3bdc7559d5afe075a51fc053e5b025534f710348: Status 404 returned error can't find the container with id 7e2b851f5b04600977b2f01f3bdc7559d5afe075a51fc053e5b025534f710348 Mar 20 09:14:51 crc kubenswrapper[4823]: I0320 09:14:51.498185 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55958644c4-vk586"] Mar 20 09:14:51 crc kubenswrapper[4823]: I0320 09:14:51.572094 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4"] Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.205432 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" event={"ID":"570a5a06-26f8-4edb-a891-703fb297adc2","Type":"ContainerStarted","Data":"34e597466673346fcd8f2bce15152f6ec68b5de1d995b72741221d8a986d28f6"} Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.205662 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.206788 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" event={"ID":"d933984b-3a5c-4792-a5cb-751cae8865ef","Type":"ContainerStarted","Data":"e97d1321bfda29443c6e805cdd7bc4c4c6efbc256e32a2c9d0f5fec608759d76"} Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.207005 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.208308 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" event={"ID":"ee7cc64f-5303-412e-83b4-322e06279025","Type":"ContainerStarted","Data":"1711664e5a363025171f8c370cb463346be080a7b43470900e17be6c5a41addb"} Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.208581 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.210298 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" event={"ID":"dfc33271-a911-4017-bb46-3d6f5ce6ddf1","Type":"ContainerStarted","Data":"5297d6a174ddb1bb260cbd3331449afe05842d7cdf404a8dbc08e9dc321a8b9b"} Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.210352 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" event={"ID":"dfc33271-a911-4017-bb46-3d6f5ce6ddf1","Type":"ContainerStarted","Data":"8c423412f9b696e6cbb0e924d3043fb46c46e70a59665ec31a1c03426092c1b2"} Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.210386 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.211453 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" event={"ID":"33c79106-0897-4b9a-ab61-5e873648f9e1","Type":"ContainerStarted","Data":"917a66532d297d2a982793426ff2bec91a7a72a3aabb53617fbdb754426767ad"} Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.211831 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.212677 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" event={"ID":"091871f2-1d5b-4263-b190-3e6262c4d1b0","Type":"ContainerStarted","Data":"f3078ec44c7f5c9f160e90c3a880e5533a7d75b72ff287707a3aed7187ddb834"} Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.214255 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" event={"ID":"daa1da6a-6c56-49b4-852d-915636736f69","Type":"ContainerStarted","Data":"7e2b851f5b04600977b2f01f3bdc7559d5afe075a51fc053e5b025534f710348"} Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.222358 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" podStartSLOduration=3.161018704 podStartE2EDuration="34.222340786s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.884652439 +0000 UTC m=+1026.570548813" lastFinishedPulling="2026-03-20 09:14:50.945974521 +0000 UTC m=+1057.631870895" observedRunningTime="2026-03-20 09:14:52.220713655 +0000 UTC m=+1058.906610029" watchObservedRunningTime="2026-03-20 09:14:52.222340786 +0000 UTC m=+1058.908237150" Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.238527 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" podStartSLOduration=3.375740293 podStartE2EDuration="34.238508055s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:20.079620358 +0000 UTC m=+1026.765516732" lastFinishedPulling="2026-03-20 09:14:50.94238812 +0000 UTC m=+1057.628284494" observedRunningTime="2026-03-20 09:14:52.235849183 +0000 UTC m=+1058.921745577" watchObservedRunningTime="2026-03-20 09:14:52.238508055 +0000 UTC m=+1058.924404429" Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.261329 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" podStartSLOduration=3.3773557690000002 podStartE2EDuration="34.261306508s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:20.066403327 +0000 UTC m=+1026.752299701" lastFinishedPulling="2026-03-20 09:14:50.950354066 +0000 UTC m=+1057.636250440" observedRunningTime="2026-03-20 09:14:52.253797906 +0000 UTC m=+1058.939694300" watchObservedRunningTime="2026-03-20 09:14:52.261306508 +0000 UTC m=+1058.947202902" Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.290919 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" podStartSLOduration=34.290903701 podStartE2EDuration="34.290903701s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:14:52.285881927 +0000 UTC m=+1058.971778301" watchObservedRunningTime="2026-03-20 09:14:52.290903701 +0000 UTC m=+1058.976800075" Mar 20 09:14:52 crc kubenswrapper[4823]: I0320 09:14:52.306234 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" podStartSLOduration=3.252783398 podStartE2EDuration="34.306217605s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.892620237 +0000 UTC m=+1026.578516611" lastFinishedPulling="2026-03-20 09:14:50.946054404 +0000 UTC m=+1057.631950818" observedRunningTime="2026-03-20 09:14:52.304589824 +0000 UTC m=+1058.990486218" watchObservedRunningTime="2026-03-20 09:14:52.306217605 +0000 UTC m=+1058.992113989" Mar 20 09:14:54 crc kubenswrapper[4823]: I0320 09:14:54.233133 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" event={"ID":"091871f2-1d5b-4263-b190-3e6262c4d1b0","Type":"ContainerStarted","Data":"77a38e6cb9d1881eafce6ec018bd5be799d5335d4b9856f862b56e2a80611da1"} Mar 20 09:14:54 crc kubenswrapper[4823]: I0320 09:14:54.234624 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:14:54 crc kubenswrapper[4823]: I0320 09:14:54.265808 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" podStartSLOduration=34.522768068 podStartE2EDuration="36.265784376s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:51.593810116 +0000 UTC m=+1058.279706490" lastFinishedPulling="2026-03-20 09:14:53.336826424 +0000 UTC m=+1060.022722798" observedRunningTime="2026-03-20 09:14:54.261188923 +0000 UTC m=+1060.947085307" watchObservedRunningTime="2026-03-20 09:14:54.265784376 +0000 UTC m=+1060.951680750" Mar 20 09:14:55 crc kubenswrapper[4823]: I0320 09:14:55.240397 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" event={"ID":"d8a80347-5759-4eb7-9750-b170fa6e5689","Type":"ContainerStarted","Data":"eee44e271c6949d2eb0a5f312e05d516ee2d5ecf7f37c3324ae851dd6474d769"} Mar 20 09:14:55 crc kubenswrapper[4823]: I0320 09:14:55.240628 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" Mar 20 09:14:55 crc kubenswrapper[4823]: I0320 09:14:55.243697 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" event={"ID":"daa1da6a-6c56-49b4-852d-915636736f69","Type":"ContainerStarted","Data":"4445ea1560f5fcdc0f88daff9f8a8a69e852cf3c4689be0dbfa060b81b41a5a7"} Mar 20 09:14:55 crc kubenswrapper[4823]: I0320 09:14:55.243754 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:14:55 crc kubenswrapper[4823]: I0320 09:14:55.276619 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" podStartSLOduration=2.918400462 podStartE2EDuration="37.276603074s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.717934829 +0000 UTC m=+1026.403831193" lastFinishedPulling="2026-03-20 09:14:54.076137431 +0000 UTC m=+1060.762033805" observedRunningTime="2026-03-20 09:14:55.260334002 +0000 UTC m=+1061.946230376" watchObservedRunningTime="2026-03-20 09:14:55.276603074 +0000 UTC m=+1061.962499448" Mar 20 09:14:58 crc kubenswrapper[4823]: I0320 09:14:58.876138 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-mbb9d" Mar 20 09:14:58 crc kubenswrapper[4823]: I0320 09:14:58.899002 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-55f864c847-xkcdj" Mar 20 09:14:58 crc kubenswrapper[4823]: I0320 09:14:58.904843 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" podStartSLOduration=38.265757825 podStartE2EDuration="40.904813467s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:51.440698361 +0000 UTC m=+1058.126594735" lastFinishedPulling="2026-03-20 09:14:54.079754003 +0000 UTC m=+1060.765650377" observedRunningTime="2026-03-20 09:14:55.275040685 +0000 UTC m=+1061.960937079" watchObservedRunningTime="2026-03-20 09:14:58.904813467 +0000 UTC m=+1065.590709861" Mar 20 09:14:58 crc kubenswrapper[4823]: I0320 09:14:58.920924 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-767865f676-8k6bt" Mar 20 09:14:59 crc kubenswrapper[4823]: I0320 09:14:59.002907 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-884679f54-fdgtg" Mar 20 09:14:59 crc kubenswrapper[4823]: I0320 09:14:59.107951 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-htq9x" Mar 20 09:14:59 crc kubenswrapper[4823]: I0320 09:14:59.143908 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-6z7z9" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.131033 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77"] Mar 20 09:15:00 crc kubenswrapper[4823]: E0320 09:15:00.131294 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" containerName="extract-utilities" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.131306 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" containerName="extract-utilities" Mar 20 09:15:00 crc kubenswrapper[4823]: E0320 09:15:00.131320 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerName="registry-server" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.131326 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerName="registry-server" Mar 20 09:15:00 crc kubenswrapper[4823]: E0320 09:15:00.131342 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" containerName="registry-server" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.131349 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" containerName="registry-server" Mar 20 09:15:00 crc kubenswrapper[4823]: E0320 09:15:00.131360 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerName="extract-utilities" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.131366 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerName="extract-utilities" Mar 20 09:15:00 crc kubenswrapper[4823]: E0320 09:15:00.131374 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" containerName="extract-content" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.131380 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" containerName="extract-content" Mar 20 09:15:00 crc kubenswrapper[4823]: E0320 09:15:00.131392 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerName="extract-content" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.131398 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerName="extract-content" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.131524 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1ce5c2e-da8c-4004-ab2a-790f563accfb" containerName="registry-server" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.131536 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e15159-c2f8-45a9-b37a-2ed8529d73c7" containerName="registry-server" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.132110 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.133843 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.136574 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.143124 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77"] Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.290099 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-config-volume\") pod \"collect-profiles-29566635-qdv77\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.290203 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrwnf\" (UniqueName: \"kubernetes.io/projected/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-kube-api-access-xrwnf\") pod \"collect-profiles-29566635-qdv77\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.290249 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-secret-volume\") pod \"collect-profiles-29566635-qdv77\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.391022 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrwnf\" (UniqueName: \"kubernetes.io/projected/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-kube-api-access-xrwnf\") pod \"collect-profiles-29566635-qdv77\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.391070 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-secret-volume\") pod \"collect-profiles-29566635-qdv77\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.391124 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-config-volume\") pod \"collect-profiles-29566635-qdv77\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.393079 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-config-volume\") pod \"collect-profiles-29566635-qdv77\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.398425 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-secret-volume\") pod \"collect-profiles-29566635-qdv77\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.408768 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrwnf\" (UniqueName: \"kubernetes.io/projected/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-kube-api-access-xrwnf\") pod \"collect-profiles-29566635-qdv77\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.450778 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.487253 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-577ccd856-ftcjs" Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.860303 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77"] Mar 20 09:15:00 crc kubenswrapper[4823]: I0320 09:15:00.862071 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-rtmw4" Mar 20 09:15:01 crc kubenswrapper[4823]: I0320 09:15:01.190137 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-vk586" Mar 20 09:15:01 crc kubenswrapper[4823]: I0320 09:15:01.289655 4823 generic.go:334] "Generic (PLEG): container finished" podID="1eee5cc6-1fab-46c8-b10d-ff8324edc1f2" containerID="c9dd3c6c61c73d4a73390b238ac82856e0dd6b7cc19e37785384823747194f61" exitCode=0 Mar 20 09:15:01 crc kubenswrapper[4823]: I0320 09:15:01.289697 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" event={"ID":"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2","Type":"ContainerDied","Data":"c9dd3c6c61c73d4a73390b238ac82856e0dd6b7cc19e37785384823747194f61"} Mar 20 09:15:01 crc kubenswrapper[4823]: I0320 09:15:01.289725 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" event={"ID":"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2","Type":"ContainerStarted","Data":"bd77d088d7e1bc8cebe720635bb6b24a5182082f1724731f47c6dbc6416d5b07"} Mar 20 09:15:02 crc kubenswrapper[4823]: I0320 09:15:02.520501 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:02 crc kubenswrapper[4823]: I0320 09:15:02.621096 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-secret-volume\") pod \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " Mar 20 09:15:02 crc kubenswrapper[4823]: I0320 09:15:02.621170 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrwnf\" (UniqueName: \"kubernetes.io/projected/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-kube-api-access-xrwnf\") pod \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " Mar 20 09:15:02 crc kubenswrapper[4823]: I0320 09:15:02.621203 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-config-volume\") pod \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\" (UID: \"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2\") " Mar 20 09:15:02 crc kubenswrapper[4823]: I0320 09:15:02.622067 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-config-volume" (OuterVolumeSpecName: "config-volume") pod "1eee5cc6-1fab-46c8-b10d-ff8324edc1f2" (UID: "1eee5cc6-1fab-46c8-b10d-ff8324edc1f2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:15:02 crc kubenswrapper[4823]: I0320 09:15:02.626798 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-kube-api-access-xrwnf" (OuterVolumeSpecName: "kube-api-access-xrwnf") pod "1eee5cc6-1fab-46c8-b10d-ff8324edc1f2" (UID: "1eee5cc6-1fab-46c8-b10d-ff8324edc1f2"). InnerVolumeSpecName "kube-api-access-xrwnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:15:02 crc kubenswrapper[4823]: I0320 09:15:02.626918 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1eee5cc6-1fab-46c8-b10d-ff8324edc1f2" (UID: "1eee5cc6-1fab-46c8-b10d-ff8324edc1f2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:15:02 crc kubenswrapper[4823]: I0320 09:15:02.722473 4823 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:15:02 crc kubenswrapper[4823]: I0320 09:15:02.722533 4823 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:15:02 crc kubenswrapper[4823]: I0320 09:15:02.722552 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrwnf\" (UniqueName: \"kubernetes.io/projected/1eee5cc6-1fab-46c8-b10d-ff8324edc1f2-kube-api-access-xrwnf\") on node \"crc\" DevicePath \"\"" Mar 20 09:15:03 crc kubenswrapper[4823]: I0320 09:15:03.310007 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" event={"ID":"1eee5cc6-1fab-46c8-b10d-ff8324edc1f2","Type":"ContainerDied","Data":"bd77d088d7e1bc8cebe720635bb6b24a5182082f1724731f47c6dbc6416d5b07"} Mar 20 09:15:03 crc kubenswrapper[4823]: I0320 09:15:03.310044 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd77d088d7e1bc8cebe720635bb6b24a5182082f1724731f47c6dbc6416d5b07" Mar 20 09:15:03 crc kubenswrapper[4823]: I0320 09:15:03.310091 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-qdv77" Mar 20 09:15:05 crc kubenswrapper[4823]: E0320 09:15:05.098405 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" podUID="638e33c0-1409-4f54-a08b-25c6ef4319a7" Mar 20 09:15:08 crc kubenswrapper[4823]: I0320 09:15:08.959654 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-28hd6" Mar 20 09:15:17 crc kubenswrapper[4823]: I0320 09:15:17.472697 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" event={"ID":"638e33c0-1409-4f54-a08b-25c6ef4319a7","Type":"ContainerStarted","Data":"21b0fc82153b9c1067a022a3fe9dddc1954bda6e369b74f5181523c6062aa09a"} Mar 20 09:15:17 crc kubenswrapper[4823]: I0320 09:15:17.473441 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" Mar 20 09:15:17 crc kubenswrapper[4823]: I0320 09:15:17.493809 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" podStartSLOduration=2.691090893 podStartE2EDuration="59.493790265s" podCreationTimestamp="2026-03-20 09:14:18 +0000 UTC" firstStartedPulling="2026-03-20 09:14:19.895311881 +0000 UTC m=+1026.581208255" lastFinishedPulling="2026-03-20 09:15:16.698011243 +0000 UTC m=+1083.383907627" observedRunningTime="2026-03-20 09:15:17.490571055 +0000 UTC m=+1084.176467429" watchObservedRunningTime="2026-03-20 09:15:17.493790265 +0000 UTC m=+1084.179686639" Mar 20 09:15:28 crc kubenswrapper[4823]: I0320 09:15:28.993713 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-vft4x" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.027169 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4dcfz"] Mar 20 09:15:46 crc kubenswrapper[4823]: E0320 09:15:46.028640 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eee5cc6-1fab-46c8-b10d-ff8324edc1f2" containerName="collect-profiles" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.031360 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eee5cc6-1fab-46c8-b10d-ff8324edc1f2" containerName="collect-profiles" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.031601 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eee5cc6-1fab-46c8-b10d-ff8324edc1f2" containerName="collect-profiles" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.032540 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.036778 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.037053 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.037111 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-hnt9w" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.037874 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.049826 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4dcfz"] Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.160481 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxt7d\" (UniqueName: \"kubernetes.io/projected/222c58e9-e172-48c2-a54d-95ea1f26c554-kube-api-access-nxt7d\") pod \"dnsmasq-dns-675f4bcbfc-4dcfz\" (UID: \"222c58e9-e172-48c2-a54d-95ea1f26c554\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.160592 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/222c58e9-e172-48c2-a54d-95ea1f26c554-config\") pod \"dnsmasq-dns-675f4bcbfc-4dcfz\" (UID: \"222c58e9-e172-48c2-a54d-95ea1f26c554\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.163865 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t9dnq"] Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.165150 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.168382 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.178061 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t9dnq"] Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.261700 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxt7d\" (UniqueName: \"kubernetes.io/projected/222c58e9-e172-48c2-a54d-95ea1f26c554-kube-api-access-nxt7d\") pod \"dnsmasq-dns-675f4bcbfc-4dcfz\" (UID: \"222c58e9-e172-48c2-a54d-95ea1f26c554\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.261803 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a2401f7-2c31-4d95-bf2c-9896d96f6ed5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-t9dnq\" (UID: \"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.261824 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x849\" (UniqueName: \"kubernetes.io/projected/5a2401f7-2c31-4d95-bf2c-9896d96f6ed5-kube-api-access-4x849\") pod \"dnsmasq-dns-78dd6ddcc-t9dnq\" (UID: \"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.261850 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a2401f7-2c31-4d95-bf2c-9896d96f6ed5-config\") pod \"dnsmasq-dns-78dd6ddcc-t9dnq\" (UID: \"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.261891 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/222c58e9-e172-48c2-a54d-95ea1f26c554-config\") pod \"dnsmasq-dns-675f4bcbfc-4dcfz\" (UID: \"222c58e9-e172-48c2-a54d-95ea1f26c554\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.262624 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/222c58e9-e172-48c2-a54d-95ea1f26c554-config\") pod \"dnsmasq-dns-675f4bcbfc-4dcfz\" (UID: \"222c58e9-e172-48c2-a54d-95ea1f26c554\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.288951 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxt7d\" (UniqueName: \"kubernetes.io/projected/222c58e9-e172-48c2-a54d-95ea1f26c554-kube-api-access-nxt7d\") pod \"dnsmasq-dns-675f4bcbfc-4dcfz\" (UID: \"222c58e9-e172-48c2-a54d-95ea1f26c554\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.354656 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.363318 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a2401f7-2c31-4d95-bf2c-9896d96f6ed5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-t9dnq\" (UID: \"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.363357 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x849\" (UniqueName: \"kubernetes.io/projected/5a2401f7-2c31-4d95-bf2c-9896d96f6ed5-kube-api-access-4x849\") pod \"dnsmasq-dns-78dd6ddcc-t9dnq\" (UID: \"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.363669 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a2401f7-2c31-4d95-bf2c-9896d96f6ed5-config\") pod \"dnsmasq-dns-78dd6ddcc-t9dnq\" (UID: \"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.364231 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a2401f7-2c31-4d95-bf2c-9896d96f6ed5-config\") pod \"dnsmasq-dns-78dd6ddcc-t9dnq\" (UID: \"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.364764 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a2401f7-2c31-4d95-bf2c-9896d96f6ed5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-t9dnq\" (UID: \"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.387379 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x849\" (UniqueName: \"kubernetes.io/projected/5a2401f7-2c31-4d95-bf2c-9896d96f6ed5-kube-api-access-4x849\") pod \"dnsmasq-dns-78dd6ddcc-t9dnq\" (UID: \"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.491539 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.782886 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4dcfz"] Mar 20 09:15:46 crc kubenswrapper[4823]: I0320 09:15:46.909494 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-t9dnq"] Mar 20 09:15:46 crc kubenswrapper[4823]: W0320 09:15:46.915697 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a2401f7_2c31_4d95_bf2c_9896d96f6ed5.slice/crio-b170e25e67d60c5c1a939a5be185f233800b4a959e455ef1cfeb546d04aa8523 WatchSource:0}: Error finding container b170e25e67d60c5c1a939a5be185f233800b4a959e455ef1cfeb546d04aa8523: Status 404 returned error can't find the container with id b170e25e67d60c5c1a939a5be185f233800b4a959e455ef1cfeb546d04aa8523 Mar 20 09:15:47 crc kubenswrapper[4823]: I0320 09:15:47.734916 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" event={"ID":"222c58e9-e172-48c2-a54d-95ea1f26c554","Type":"ContainerStarted","Data":"41e293d76b94554124c728f976c9b7b18000f8f0a0c5920ec89aaa219858518e"} Mar 20 09:15:47 crc kubenswrapper[4823]: I0320 09:15:47.736326 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" event={"ID":"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5","Type":"ContainerStarted","Data":"b170e25e67d60c5c1a939a5be185f233800b4a959e455ef1cfeb546d04aa8523"} Mar 20 09:16:00 crc kubenswrapper[4823]: I0320 09:16:00.128977 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566636-xtzck"] Mar 20 09:16:00 crc kubenswrapper[4823]: I0320 09:16:00.135336 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566636-xtzck" Mar 20 09:16:00 crc kubenswrapper[4823]: I0320 09:16:00.140987 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:16:00 crc kubenswrapper[4823]: I0320 09:16:00.141779 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:16:00 crc kubenswrapper[4823]: I0320 09:16:00.142018 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:16:00 crc kubenswrapper[4823]: I0320 09:16:00.171981 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566636-xtzck"] Mar 20 09:16:00 crc kubenswrapper[4823]: I0320 09:16:00.247480 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4822\" (UniqueName: \"kubernetes.io/projected/5cf48399-a918-4f54-89b2-0891d06d043b-kube-api-access-v4822\") pod \"auto-csr-approver-29566636-xtzck\" (UID: \"5cf48399-a918-4f54-89b2-0891d06d043b\") " pod="openshift-infra/auto-csr-approver-29566636-xtzck" Mar 20 09:16:00 crc kubenswrapper[4823]: I0320 09:16:00.349145 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4822\" (UniqueName: \"kubernetes.io/projected/5cf48399-a918-4f54-89b2-0891d06d043b-kube-api-access-v4822\") pod \"auto-csr-approver-29566636-xtzck\" (UID: \"5cf48399-a918-4f54-89b2-0891d06d043b\") " pod="openshift-infra/auto-csr-approver-29566636-xtzck" Mar 20 09:16:00 crc kubenswrapper[4823]: I0320 09:16:00.373683 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4822\" (UniqueName: \"kubernetes.io/projected/5cf48399-a918-4f54-89b2-0891d06d043b-kube-api-access-v4822\") pod \"auto-csr-approver-29566636-xtzck\" (UID: \"5cf48399-a918-4f54-89b2-0891d06d043b\") " pod="openshift-infra/auto-csr-approver-29566636-xtzck" Mar 20 09:16:00 crc kubenswrapper[4823]: I0320 09:16:00.472533 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566636-xtzck" Mar 20 09:16:02 crc kubenswrapper[4823]: E0320 09:16:02.595358 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 20 09:16:02 crc kubenswrapper[4823]: E0320 09:16:02.595860 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nxt7d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-4dcfz_openstack(222c58e9-e172-48c2-a54d-95ea1f26c554): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:16:02 crc kubenswrapper[4823]: E0320 09:16:02.597059 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" podUID="222c58e9-e172-48c2-a54d-95ea1f26c554" Mar 20 09:16:02 crc kubenswrapper[4823]: E0320 09:16:02.599955 4823 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 20 09:16:02 crc kubenswrapper[4823]: E0320 09:16:02.600124 4823 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4x849,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-t9dnq_openstack(5a2401f7-2c31-4d95-bf2c-9896d96f6ed5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:16:02 crc kubenswrapper[4823]: E0320 09:16:02.601282 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" podUID="5a2401f7-2c31-4d95-bf2c-9896d96f6ed5" Mar 20 09:16:02 crc kubenswrapper[4823]: E0320 09:16:02.859102 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" podUID="222c58e9-e172-48c2-a54d-95ea1f26c554" Mar 20 09:16:02 crc kubenswrapper[4823]: E0320 09:16:02.859185 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" podUID="5a2401f7-2c31-4d95-bf2c-9896d96f6ed5" Mar 20 09:16:02 crc kubenswrapper[4823]: I0320 09:16:02.954569 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566636-xtzck"] Mar 20 09:16:02 crc kubenswrapper[4823]: W0320 09:16:02.960414 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cf48399_a918_4f54_89b2_0891d06d043b.slice/crio-45717632c6cf091496c7967c259ac29e82893e4ead251677f6bb2c8f5dcdc044 WatchSource:0}: Error finding container 45717632c6cf091496c7967c259ac29e82893e4ead251677f6bb2c8f5dcdc044: Status 404 returned error can't find the container with id 45717632c6cf091496c7967c259ac29e82893e4ead251677f6bb2c8f5dcdc044 Mar 20 09:16:02 crc kubenswrapper[4823]: I0320 09:16:02.963702 4823 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:16:03 crc kubenswrapper[4823]: I0320 09:16:03.864480 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566636-xtzck" event={"ID":"5cf48399-a918-4f54-89b2-0891d06d043b","Type":"ContainerStarted","Data":"45717632c6cf091496c7967c259ac29e82893e4ead251677f6bb2c8f5dcdc044"} Mar 20 09:16:04 crc kubenswrapper[4823]: I0320 09:16:04.875135 4823 generic.go:334] "Generic (PLEG): container finished" podID="5cf48399-a918-4f54-89b2-0891d06d043b" containerID="0b249db9d3205e23cb180145a20b04e976969d5ccb423d493b6bed25f85cea37" exitCode=0 Mar 20 09:16:04 crc kubenswrapper[4823]: I0320 09:16:04.875204 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566636-xtzck" event={"ID":"5cf48399-a918-4f54-89b2-0891d06d043b","Type":"ContainerDied","Data":"0b249db9d3205e23cb180145a20b04e976969d5ccb423d493b6bed25f85cea37"} Mar 20 09:16:06 crc kubenswrapper[4823]: I0320 09:16:06.186459 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566636-xtzck" Mar 20 09:16:06 crc kubenswrapper[4823]: I0320 09:16:06.345809 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4822\" (UniqueName: \"kubernetes.io/projected/5cf48399-a918-4f54-89b2-0891d06d043b-kube-api-access-v4822\") pod \"5cf48399-a918-4f54-89b2-0891d06d043b\" (UID: \"5cf48399-a918-4f54-89b2-0891d06d043b\") " Mar 20 09:16:06 crc kubenswrapper[4823]: I0320 09:16:06.353932 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cf48399-a918-4f54-89b2-0891d06d043b-kube-api-access-v4822" (OuterVolumeSpecName: "kube-api-access-v4822") pod "5cf48399-a918-4f54-89b2-0891d06d043b" (UID: "5cf48399-a918-4f54-89b2-0891d06d043b"). InnerVolumeSpecName "kube-api-access-v4822". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:16:06 crc kubenswrapper[4823]: I0320 09:16:06.447678 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4822\" (UniqueName: \"kubernetes.io/projected/5cf48399-a918-4f54-89b2-0891d06d043b-kube-api-access-v4822\") on node \"crc\" DevicePath \"\"" Mar 20 09:16:06 crc kubenswrapper[4823]: I0320 09:16:06.891446 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566636-xtzck" event={"ID":"5cf48399-a918-4f54-89b2-0891d06d043b","Type":"ContainerDied","Data":"45717632c6cf091496c7967c259ac29e82893e4ead251677f6bb2c8f5dcdc044"} Mar 20 09:16:06 crc kubenswrapper[4823]: I0320 09:16:06.891495 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45717632c6cf091496c7967c259ac29e82893e4ead251677f6bb2c8f5dcdc044" Mar 20 09:16:06 crc kubenswrapper[4823]: I0320 09:16:06.891577 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566636-xtzck" Mar 20 09:16:07 crc kubenswrapper[4823]: I0320 09:16:07.256094 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566630-r54tb"] Mar 20 09:16:07 crc kubenswrapper[4823]: I0320 09:16:07.260583 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566630-r54tb"] Mar 20 09:16:08 crc kubenswrapper[4823]: I0320 09:16:08.108937 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3431f35-1655-4dae-89d3-d30c2a5065bb" path="/var/lib/kubelet/pods/e3431f35-1655-4dae-89d3-d30c2a5065bb/volumes" Mar 20 09:16:09 crc kubenswrapper[4823]: I0320 09:16:09.430642 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:16:09 crc kubenswrapper[4823]: I0320 09:16:09.430693 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:16:15 crc kubenswrapper[4823]: I0320 09:16:15.963439 4823 generic.go:334] "Generic (PLEG): container finished" podID="222c58e9-e172-48c2-a54d-95ea1f26c554" containerID="ff82c006f8303d924f5c6f4b9437c3c75cb87bd35533f7d48750c03e9b047acd" exitCode=0 Mar 20 09:16:15 crc kubenswrapper[4823]: I0320 09:16:15.963578 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" event={"ID":"222c58e9-e172-48c2-a54d-95ea1f26c554","Type":"ContainerDied","Data":"ff82c006f8303d924f5c6f4b9437c3c75cb87bd35533f7d48750c03e9b047acd"} Mar 20 09:16:16 crc kubenswrapper[4823]: I0320 09:16:16.978027 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" event={"ID":"222c58e9-e172-48c2-a54d-95ea1f26c554","Type":"ContainerStarted","Data":"be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e"} Mar 20 09:16:16 crc kubenswrapper[4823]: I0320 09:16:16.978632 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:16:17 crc kubenswrapper[4823]: I0320 09:16:17.006962 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" podStartSLOduration=2.221852738 podStartE2EDuration="31.006927694s" podCreationTimestamp="2026-03-20 09:15:46 +0000 UTC" firstStartedPulling="2026-03-20 09:15:46.788650783 +0000 UTC m=+1113.474547157" lastFinishedPulling="2026-03-20 09:16:15.573725709 +0000 UTC m=+1142.259622113" observedRunningTime="2026-03-20 09:16:17.004768428 +0000 UTC m=+1143.690664802" watchObservedRunningTime="2026-03-20 09:16:17.006927694 +0000 UTC m=+1143.692824068" Mar 20 09:16:18 crc kubenswrapper[4823]: I0320 09:16:18.993151 4823 generic.go:334] "Generic (PLEG): container finished" podID="5a2401f7-2c31-4d95-bf2c-9896d96f6ed5" containerID="bb7ee10af1c8f37a52f0c06b3d57d76d1f0a12d006625780a80ddb54a63bacc7" exitCode=0 Mar 20 09:16:18 crc kubenswrapper[4823]: I0320 09:16:18.993230 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" event={"ID":"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5","Type":"ContainerDied","Data":"bb7ee10af1c8f37a52f0c06b3d57d76d1f0a12d006625780a80ddb54a63bacc7"} Mar 20 09:16:20 crc kubenswrapper[4823]: I0320 09:16:20.002927 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" event={"ID":"5a2401f7-2c31-4d95-bf2c-9896d96f6ed5","Type":"ContainerStarted","Data":"d583572f741dbe39a63a6a966097c1c12261641b155458d82a6afc0767cb927b"} Mar 20 09:16:20 crc kubenswrapper[4823]: I0320 09:16:20.003498 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:16:20 crc kubenswrapper[4823]: I0320 09:16:20.022736 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" podStartSLOduration=-9223372002.832058 podStartE2EDuration="34.022718085s" podCreationTimestamp="2026-03-20 09:15:46 +0000 UTC" firstStartedPulling="2026-03-20 09:15:46.918770969 +0000 UTC m=+1113.604667343" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:16:20.019136474 +0000 UTC m=+1146.705032848" watchObservedRunningTime="2026-03-20 09:16:20.022718085 +0000 UTC m=+1146.708614459" Mar 20 09:16:21 crc kubenswrapper[4823]: I0320 09:16:21.356957 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:16:26 crc kubenswrapper[4823]: I0320 09:16:26.492952 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78dd6ddcc-t9dnq" Mar 20 09:16:26 crc kubenswrapper[4823]: I0320 09:16:26.549181 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4dcfz"] Mar 20 09:16:26 crc kubenswrapper[4823]: I0320 09:16:26.549446 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" podUID="222c58e9-e172-48c2-a54d-95ea1f26c554" containerName="dnsmasq-dns" containerID="cri-o://be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e" gracePeriod=10 Mar 20 09:16:26 crc kubenswrapper[4823]: I0320 09:16:26.927483 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:16:26 crc kubenswrapper[4823]: I0320 09:16:26.995201 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/222c58e9-e172-48c2-a54d-95ea1f26c554-config\") pod \"222c58e9-e172-48c2-a54d-95ea1f26c554\" (UID: \"222c58e9-e172-48c2-a54d-95ea1f26c554\") " Mar 20 09:16:26 crc kubenswrapper[4823]: I0320 09:16:26.995320 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxt7d\" (UniqueName: \"kubernetes.io/projected/222c58e9-e172-48c2-a54d-95ea1f26c554-kube-api-access-nxt7d\") pod \"222c58e9-e172-48c2-a54d-95ea1f26c554\" (UID: \"222c58e9-e172-48c2-a54d-95ea1f26c554\") " Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.006913 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/222c58e9-e172-48c2-a54d-95ea1f26c554-kube-api-access-nxt7d" (OuterVolumeSpecName: "kube-api-access-nxt7d") pod "222c58e9-e172-48c2-a54d-95ea1f26c554" (UID: "222c58e9-e172-48c2-a54d-95ea1f26c554"). InnerVolumeSpecName "kube-api-access-nxt7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.070387 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/222c58e9-e172-48c2-a54d-95ea1f26c554-config" (OuterVolumeSpecName: "config") pod "222c58e9-e172-48c2-a54d-95ea1f26c554" (UID: "222c58e9-e172-48c2-a54d-95ea1f26c554"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.096588 4823 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/222c58e9-e172-48c2-a54d-95ea1f26c554-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.096754 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxt7d\" (UniqueName: \"kubernetes.io/projected/222c58e9-e172-48c2-a54d-95ea1f26c554-kube-api-access-nxt7d\") on node \"crc\" DevicePath \"\"" Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.591408 4823 generic.go:334] "Generic (PLEG): container finished" podID="222c58e9-e172-48c2-a54d-95ea1f26c554" containerID="be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e" exitCode=0 Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.591468 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" event={"ID":"222c58e9-e172-48c2-a54d-95ea1f26c554","Type":"ContainerDied","Data":"be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e"} Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.591521 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.591551 4823 scope.go:117] "RemoveContainer" containerID="be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e" Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.591530 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4dcfz" event={"ID":"222c58e9-e172-48c2-a54d-95ea1f26c554","Type":"ContainerDied","Data":"41e293d76b94554124c728f976c9b7b18000f8f0a0c5920ec89aaa219858518e"} Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.611302 4823 scope.go:117] "RemoveContainer" containerID="ff82c006f8303d924f5c6f4b9437c3c75cb87bd35533f7d48750c03e9b047acd" Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.621637 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4dcfz"] Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.627509 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4dcfz"] Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.639579 4823 scope.go:117] "RemoveContainer" containerID="be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e" Mar 20 09:16:27 crc kubenswrapper[4823]: E0320 09:16:27.640065 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e\": container with ID starting with be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e not found: ID does not exist" containerID="be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e" Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.640100 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e"} err="failed to get container status \"be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e\": rpc error: code = NotFound desc = could not find container \"be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e\": container with ID starting with be8510b487f2b392727bad745164c6bdb23d1d2fd37a11e06adc516907f6ee1e not found: ID does not exist" Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.640124 4823 scope.go:117] "RemoveContainer" containerID="ff82c006f8303d924f5c6f4b9437c3c75cb87bd35533f7d48750c03e9b047acd" Mar 20 09:16:27 crc kubenswrapper[4823]: E0320 09:16:27.640583 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff82c006f8303d924f5c6f4b9437c3c75cb87bd35533f7d48750c03e9b047acd\": container with ID starting with ff82c006f8303d924f5c6f4b9437c3c75cb87bd35533f7d48750c03e9b047acd not found: ID does not exist" containerID="ff82c006f8303d924f5c6f4b9437c3c75cb87bd35533f7d48750c03e9b047acd" Mar 20 09:16:27 crc kubenswrapper[4823]: I0320 09:16:27.640734 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff82c006f8303d924f5c6f4b9437c3c75cb87bd35533f7d48750c03e9b047acd"} err="failed to get container status \"ff82c006f8303d924f5c6f4b9437c3c75cb87bd35533f7d48750c03e9b047acd\": rpc error: code = NotFound desc = could not find container \"ff82c006f8303d924f5c6f4b9437c3c75cb87bd35533f7d48750c03e9b047acd\": container with ID starting with ff82c006f8303d924f5c6f4b9437c3c75cb87bd35533f7d48750c03e9b047acd not found: ID does not exist" Mar 20 09:16:28 crc kubenswrapper[4823]: I0320 09:16:28.107080 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="222c58e9-e172-48c2-a54d-95ea1f26c554" path="/var/lib/kubelet/pods/222c58e9-e172-48c2-a54d-95ea1f26c554/volumes" Mar 20 09:16:38 crc kubenswrapper[4823]: I0320 09:16:38.583534 4823 scope.go:117] "RemoveContainer" containerID="e4023c35ef355eac2c6cc566acf642f1f05b82742e42861f1a78fd51ff85e10d" Mar 20 09:16:39 crc kubenswrapper[4823]: I0320 09:16:39.430061 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:16:39 crc kubenswrapper[4823]: I0320 09:16:39.430637 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:17:09 crc kubenswrapper[4823]: I0320 09:17:09.430040 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:17:09 crc kubenswrapper[4823]: I0320 09:17:09.430823 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:17:09 crc kubenswrapper[4823]: I0320 09:17:09.430886 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:17:09 crc kubenswrapper[4823]: I0320 09:17:09.431879 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b8674863c83970000faa79c02318a604d3777e9ca2e37590494e528f8d2cbc7"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:17:09 crc kubenswrapper[4823]: I0320 09:17:09.432046 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://0b8674863c83970000faa79c02318a604d3777e9ca2e37590494e528f8d2cbc7" gracePeriod=600 Mar 20 09:17:10 crc kubenswrapper[4823]: I0320 09:17:10.194824 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="0b8674863c83970000faa79c02318a604d3777e9ca2e37590494e528f8d2cbc7" exitCode=0 Mar 20 09:17:10 crc kubenswrapper[4823]: I0320 09:17:10.194831 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"0b8674863c83970000faa79c02318a604d3777e9ca2e37590494e528f8d2cbc7"} Mar 20 09:17:10 crc kubenswrapper[4823]: I0320 09:17:10.195502 4823 scope.go:117] "RemoveContainer" containerID="3c6623b862df4196951c10211c83ea8d06f08a5d6875319545662d0ea4663061" Mar 20 09:17:11 crc kubenswrapper[4823]: I0320 09:17:11.206545 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"78975c9ca0024f8f8c44d3ccb13c9eb4809cc513d430445cccc664f308a39c85"} Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.138188 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566638-vf9vh"] Mar 20 09:18:00 crc kubenswrapper[4823]: E0320 09:18:00.139260 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222c58e9-e172-48c2-a54d-95ea1f26c554" containerName="init" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.139292 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="222c58e9-e172-48c2-a54d-95ea1f26c554" containerName="init" Mar 20 09:18:00 crc kubenswrapper[4823]: E0320 09:18:00.139303 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf48399-a918-4f54-89b2-0891d06d043b" containerName="oc" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.139311 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf48399-a918-4f54-89b2-0891d06d043b" containerName="oc" Mar 20 09:18:00 crc kubenswrapper[4823]: E0320 09:18:00.139329 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222c58e9-e172-48c2-a54d-95ea1f26c554" containerName="dnsmasq-dns" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.139338 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="222c58e9-e172-48c2-a54d-95ea1f26c554" containerName="dnsmasq-dns" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.139491 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="222c58e9-e172-48c2-a54d-95ea1f26c554" containerName="dnsmasq-dns" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.139510 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cf48399-a918-4f54-89b2-0891d06d043b" containerName="oc" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.140087 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566638-vf9vh" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.143243 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.143262 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.143419 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.146088 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566638-vf9vh"] Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.239197 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm772\" (UniqueName: \"kubernetes.io/projected/834573f2-b90a-48f0-97ab-f5cff6141a90-kube-api-access-wm772\") pod \"auto-csr-approver-29566638-vf9vh\" (UID: \"834573f2-b90a-48f0-97ab-f5cff6141a90\") " pod="openshift-infra/auto-csr-approver-29566638-vf9vh" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.340904 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm772\" (UniqueName: \"kubernetes.io/projected/834573f2-b90a-48f0-97ab-f5cff6141a90-kube-api-access-wm772\") pod \"auto-csr-approver-29566638-vf9vh\" (UID: \"834573f2-b90a-48f0-97ab-f5cff6141a90\") " pod="openshift-infra/auto-csr-approver-29566638-vf9vh" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.366980 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm772\" (UniqueName: \"kubernetes.io/projected/834573f2-b90a-48f0-97ab-f5cff6141a90-kube-api-access-wm772\") pod \"auto-csr-approver-29566638-vf9vh\" (UID: \"834573f2-b90a-48f0-97ab-f5cff6141a90\") " pod="openshift-infra/auto-csr-approver-29566638-vf9vh" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.463045 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566638-vf9vh" Mar 20 09:18:00 crc kubenswrapper[4823]: I0320 09:18:00.864637 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566638-vf9vh"] Mar 20 09:18:01 crc kubenswrapper[4823]: I0320 09:18:01.572113 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566638-vf9vh" event={"ID":"834573f2-b90a-48f0-97ab-f5cff6141a90","Type":"ContainerStarted","Data":"2c1e4f4935ea802afaa34d8078a82bea3ee0600fc7d6a2fa91b2246b30e70800"} Mar 20 09:18:02 crc kubenswrapper[4823]: I0320 09:18:02.579813 4823 generic.go:334] "Generic (PLEG): container finished" podID="834573f2-b90a-48f0-97ab-f5cff6141a90" containerID="79014211ee87220c7250d3e5d4753caf8f54863a176aeaf51bedbe4a5ec4fc6a" exitCode=0 Mar 20 09:18:02 crc kubenswrapper[4823]: I0320 09:18:02.579916 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566638-vf9vh" event={"ID":"834573f2-b90a-48f0-97ab-f5cff6141a90","Type":"ContainerDied","Data":"79014211ee87220c7250d3e5d4753caf8f54863a176aeaf51bedbe4a5ec4fc6a"} Mar 20 09:18:03 crc kubenswrapper[4823]: I0320 09:18:03.866933 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566638-vf9vh" Mar 20 09:18:03 crc kubenswrapper[4823]: I0320 09:18:03.991632 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm772\" (UniqueName: \"kubernetes.io/projected/834573f2-b90a-48f0-97ab-f5cff6141a90-kube-api-access-wm772\") pod \"834573f2-b90a-48f0-97ab-f5cff6141a90\" (UID: \"834573f2-b90a-48f0-97ab-f5cff6141a90\") " Mar 20 09:18:03 crc kubenswrapper[4823]: I0320 09:18:03.997242 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/834573f2-b90a-48f0-97ab-f5cff6141a90-kube-api-access-wm772" (OuterVolumeSpecName: "kube-api-access-wm772") pod "834573f2-b90a-48f0-97ab-f5cff6141a90" (UID: "834573f2-b90a-48f0-97ab-f5cff6141a90"). InnerVolumeSpecName "kube-api-access-wm772". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:18:04 crc kubenswrapper[4823]: I0320 09:18:04.093197 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm772\" (UniqueName: \"kubernetes.io/projected/834573f2-b90a-48f0-97ab-f5cff6141a90-kube-api-access-wm772\") on node \"crc\" DevicePath \"\"" Mar 20 09:18:04 crc kubenswrapper[4823]: I0320 09:18:04.594438 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566638-vf9vh" event={"ID":"834573f2-b90a-48f0-97ab-f5cff6141a90","Type":"ContainerDied","Data":"2c1e4f4935ea802afaa34d8078a82bea3ee0600fc7d6a2fa91b2246b30e70800"} Mar 20 09:18:04 crc kubenswrapper[4823]: I0320 09:18:04.594479 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c1e4f4935ea802afaa34d8078a82bea3ee0600fc7d6a2fa91b2246b30e70800" Mar 20 09:18:04 crc kubenswrapper[4823]: I0320 09:18:04.594546 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566638-vf9vh" Mar 20 09:18:04 crc kubenswrapper[4823]: I0320 09:18:04.930557 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566632-6mjng"] Mar 20 09:18:04 crc kubenswrapper[4823]: I0320 09:18:04.935163 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566632-6mjng"] Mar 20 09:18:06 crc kubenswrapper[4823]: I0320 09:18:06.108950 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e23c5c6-be76-47f5-baf3-8cfa530dea27" path="/var/lib/kubelet/pods/3e23c5c6-be76-47f5-baf3-8cfa530dea27/volumes" Mar 20 09:18:38 crc kubenswrapper[4823]: I0320 09:18:38.670951 4823 scope.go:117] "RemoveContainer" containerID="bb63e63c35cd3a631abe26756abd61ed410a2fef3ce422b6c2b9c8dcea0a1b0b" Mar 20 09:19:39 crc kubenswrapper[4823]: I0320 09:19:39.430886 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:19:39 crc kubenswrapper[4823]: I0320 09:19:39.431709 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.143970 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566640-5phwr"] Mar 20 09:20:00 crc kubenswrapper[4823]: E0320 09:20:00.145010 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="834573f2-b90a-48f0-97ab-f5cff6141a90" containerName="oc" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.145027 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="834573f2-b90a-48f0-97ab-f5cff6141a90" containerName="oc" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.145204 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="834573f2-b90a-48f0-97ab-f5cff6141a90" containerName="oc" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.145757 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566640-5phwr" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.147882 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.147980 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.148053 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.157727 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566640-5phwr"] Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.254645 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjq2t\" (UniqueName: \"kubernetes.io/projected/fac8eee5-c573-4911-b56b-f4804a790292-kube-api-access-wjq2t\") pod \"auto-csr-approver-29566640-5phwr\" (UID: \"fac8eee5-c573-4911-b56b-f4804a790292\") " pod="openshift-infra/auto-csr-approver-29566640-5phwr" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.355967 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjq2t\" (UniqueName: \"kubernetes.io/projected/fac8eee5-c573-4911-b56b-f4804a790292-kube-api-access-wjq2t\") pod \"auto-csr-approver-29566640-5phwr\" (UID: \"fac8eee5-c573-4911-b56b-f4804a790292\") " pod="openshift-infra/auto-csr-approver-29566640-5phwr" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.375280 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjq2t\" (UniqueName: \"kubernetes.io/projected/fac8eee5-c573-4911-b56b-f4804a790292-kube-api-access-wjq2t\") pod \"auto-csr-approver-29566640-5phwr\" (UID: \"fac8eee5-c573-4911-b56b-f4804a790292\") " pod="openshift-infra/auto-csr-approver-29566640-5phwr" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.466285 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566640-5phwr" Mar 20 09:20:00 crc kubenswrapper[4823]: I0320 09:20:00.887950 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566640-5phwr"] Mar 20 09:20:01 crc kubenswrapper[4823]: I0320 09:20:01.510120 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566640-5phwr" event={"ID":"fac8eee5-c573-4911-b56b-f4804a790292","Type":"ContainerStarted","Data":"c65f1616d1e6389355e9dbdabcf6c5b1c317689ae2b14eba7a7dac578e386ff3"} Mar 20 09:20:02 crc kubenswrapper[4823]: I0320 09:20:02.520731 4823 generic.go:334] "Generic (PLEG): container finished" podID="fac8eee5-c573-4911-b56b-f4804a790292" containerID="e77a907b22b846213e43d0cd198abd38f417f42e49737282b2d9e0a187a5551c" exitCode=0 Mar 20 09:20:02 crc kubenswrapper[4823]: I0320 09:20:02.520837 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566640-5phwr" event={"ID":"fac8eee5-c573-4911-b56b-f4804a790292","Type":"ContainerDied","Data":"e77a907b22b846213e43d0cd198abd38f417f42e49737282b2d9e0a187a5551c"} Mar 20 09:20:03 crc kubenswrapper[4823]: I0320 09:20:03.792348 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566640-5phwr" Mar 20 09:20:03 crc kubenswrapper[4823]: I0320 09:20:03.904045 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjq2t\" (UniqueName: \"kubernetes.io/projected/fac8eee5-c573-4911-b56b-f4804a790292-kube-api-access-wjq2t\") pod \"fac8eee5-c573-4911-b56b-f4804a790292\" (UID: \"fac8eee5-c573-4911-b56b-f4804a790292\") " Mar 20 09:20:03 crc kubenswrapper[4823]: I0320 09:20:03.910701 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fac8eee5-c573-4911-b56b-f4804a790292-kube-api-access-wjq2t" (OuterVolumeSpecName: "kube-api-access-wjq2t") pod "fac8eee5-c573-4911-b56b-f4804a790292" (UID: "fac8eee5-c573-4911-b56b-f4804a790292"). InnerVolumeSpecName "kube-api-access-wjq2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:20:04 crc kubenswrapper[4823]: I0320 09:20:04.005409 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjq2t\" (UniqueName: \"kubernetes.io/projected/fac8eee5-c573-4911-b56b-f4804a790292-kube-api-access-wjq2t\") on node \"crc\" DevicePath \"\"" Mar 20 09:20:04 crc kubenswrapper[4823]: I0320 09:20:04.537433 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566640-5phwr" event={"ID":"fac8eee5-c573-4911-b56b-f4804a790292","Type":"ContainerDied","Data":"c65f1616d1e6389355e9dbdabcf6c5b1c317689ae2b14eba7a7dac578e386ff3"} Mar 20 09:20:04 crc kubenswrapper[4823]: I0320 09:20:04.537481 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c65f1616d1e6389355e9dbdabcf6c5b1c317689ae2b14eba7a7dac578e386ff3" Mar 20 09:20:04 crc kubenswrapper[4823]: I0320 09:20:04.537546 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566640-5phwr" Mar 20 09:20:04 crc kubenswrapper[4823]: I0320 09:20:04.860259 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566634-78xrx"] Mar 20 09:20:04 crc kubenswrapper[4823]: I0320 09:20:04.866723 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566634-78xrx"] Mar 20 09:20:06 crc kubenswrapper[4823]: I0320 09:20:06.105790 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37e5ce93-ad64-4864-9430-45fb877fd038" path="/var/lib/kubelet/pods/37e5ce93-ad64-4864-9430-45fb877fd038/volumes" Mar 20 09:20:09 crc kubenswrapper[4823]: I0320 09:20:09.430005 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:20:09 crc kubenswrapper[4823]: I0320 09:20:09.430479 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:20:38 crc kubenswrapper[4823]: I0320 09:20:38.743097 4823 scope.go:117] "RemoveContainer" containerID="4378b83304dcb3dd2ae8300b75d65d9c7f24a2a9af713826305396acafa0651d" Mar 20 09:20:39 crc kubenswrapper[4823]: I0320 09:20:39.430560 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:20:39 crc kubenswrapper[4823]: I0320 09:20:39.430625 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:20:39 crc kubenswrapper[4823]: I0320 09:20:39.430750 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:20:39 crc kubenswrapper[4823]: I0320 09:20:39.431190 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"78975c9ca0024f8f8c44d3ccb13c9eb4809cc513d430445cccc664f308a39c85"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:20:39 crc kubenswrapper[4823]: I0320 09:20:39.431248 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://78975c9ca0024f8f8c44d3ccb13c9eb4809cc513d430445cccc664f308a39c85" gracePeriod=600 Mar 20 09:20:39 crc kubenswrapper[4823]: I0320 09:20:39.814634 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="78975c9ca0024f8f8c44d3ccb13c9eb4809cc513d430445cccc664f308a39c85" exitCode=0 Mar 20 09:20:39 crc kubenswrapper[4823]: I0320 09:20:39.814664 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"78975c9ca0024f8f8c44d3ccb13c9eb4809cc513d430445cccc664f308a39c85"} Mar 20 09:20:39 crc kubenswrapper[4823]: I0320 09:20:39.815103 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8"} Mar 20 09:20:39 crc kubenswrapper[4823]: I0320 09:20:39.815128 4823 scope.go:117] "RemoveContainer" containerID="0b8674863c83970000faa79c02318a604d3777e9ca2e37590494e528f8d2cbc7" Mar 20 09:21:12 crc kubenswrapper[4823]: I0320 09:21:12.865358 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2rb9v"] Mar 20 09:21:12 crc kubenswrapper[4823]: E0320 09:21:12.866898 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac8eee5-c573-4911-b56b-f4804a790292" containerName="oc" Mar 20 09:21:12 crc kubenswrapper[4823]: I0320 09:21:12.866929 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac8eee5-c573-4911-b56b-f4804a790292" containerName="oc" Mar 20 09:21:12 crc kubenswrapper[4823]: I0320 09:21:12.867377 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac8eee5-c573-4911-b56b-f4804a790292" containerName="oc" Mar 20 09:21:12 crc kubenswrapper[4823]: I0320 09:21:12.869928 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:12 crc kubenswrapper[4823]: I0320 09:21:12.879310 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2rb9v"] Mar 20 09:21:13 crc kubenswrapper[4823]: I0320 09:21:13.000400 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jddg\" (UniqueName: \"kubernetes.io/projected/4983947e-6154-4664-a5f0-9b5631076517-kube-api-access-4jddg\") pod \"redhat-operators-2rb9v\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:13 crc kubenswrapper[4823]: I0320 09:21:13.000451 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-utilities\") pod \"redhat-operators-2rb9v\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:13 crc kubenswrapper[4823]: I0320 09:21:13.000484 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-catalog-content\") pod \"redhat-operators-2rb9v\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:13 crc kubenswrapper[4823]: I0320 09:21:13.101712 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jddg\" (UniqueName: \"kubernetes.io/projected/4983947e-6154-4664-a5f0-9b5631076517-kube-api-access-4jddg\") pod \"redhat-operators-2rb9v\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:13 crc kubenswrapper[4823]: I0320 09:21:13.101777 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-utilities\") pod \"redhat-operators-2rb9v\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:13 crc kubenswrapper[4823]: I0320 09:21:13.101808 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-catalog-content\") pod \"redhat-operators-2rb9v\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:13 crc kubenswrapper[4823]: I0320 09:21:13.102319 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-catalog-content\") pod \"redhat-operators-2rb9v\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:13 crc kubenswrapper[4823]: I0320 09:21:13.102478 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-utilities\") pod \"redhat-operators-2rb9v\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:13 crc kubenswrapper[4823]: I0320 09:21:13.121663 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jddg\" (UniqueName: \"kubernetes.io/projected/4983947e-6154-4664-a5f0-9b5631076517-kube-api-access-4jddg\") pod \"redhat-operators-2rb9v\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:13 crc kubenswrapper[4823]: I0320 09:21:13.204619 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:13 crc kubenswrapper[4823]: I0320 09:21:13.628375 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2rb9v"] Mar 20 09:21:13 crc kubenswrapper[4823]: W0320 09:21:13.636619 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4983947e_6154_4664_a5f0_9b5631076517.slice/crio-ccdc6ec5a38ea640b9e619f3f4c52d2a74f2aa78fca0e239dfe7ce0e9a25d0c7 WatchSource:0}: Error finding container ccdc6ec5a38ea640b9e619f3f4c52d2a74f2aa78fca0e239dfe7ce0e9a25d0c7: Status 404 returned error can't find the container with id ccdc6ec5a38ea640b9e619f3f4c52d2a74f2aa78fca0e239dfe7ce0e9a25d0c7 Mar 20 09:21:14 crc kubenswrapper[4823]: I0320 09:21:14.505554 4823 generic.go:334] "Generic (PLEG): container finished" podID="4983947e-6154-4664-a5f0-9b5631076517" containerID="dc7206c2c104cd400cb12bb47a9f58bc6cb68291cf9e14dad33ff503a653061e" exitCode=0 Mar 20 09:21:14 crc kubenswrapper[4823]: I0320 09:21:14.505609 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rb9v" event={"ID":"4983947e-6154-4664-a5f0-9b5631076517","Type":"ContainerDied","Data":"dc7206c2c104cd400cb12bb47a9f58bc6cb68291cf9e14dad33ff503a653061e"} Mar 20 09:21:14 crc kubenswrapper[4823]: I0320 09:21:14.505641 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rb9v" event={"ID":"4983947e-6154-4664-a5f0-9b5631076517","Type":"ContainerStarted","Data":"ccdc6ec5a38ea640b9e619f3f4c52d2a74f2aa78fca0e239dfe7ce0e9a25d0c7"} Mar 20 09:21:14 crc kubenswrapper[4823]: I0320 09:21:14.507726 4823 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:21:15 crc kubenswrapper[4823]: I0320 09:21:15.520675 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rb9v" event={"ID":"4983947e-6154-4664-a5f0-9b5631076517","Type":"ContainerStarted","Data":"2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1"} Mar 20 09:21:16 crc kubenswrapper[4823]: I0320 09:21:16.529334 4823 generic.go:334] "Generic (PLEG): container finished" podID="4983947e-6154-4664-a5f0-9b5631076517" containerID="2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1" exitCode=0 Mar 20 09:21:16 crc kubenswrapper[4823]: I0320 09:21:16.529369 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rb9v" event={"ID":"4983947e-6154-4664-a5f0-9b5631076517","Type":"ContainerDied","Data":"2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1"} Mar 20 09:21:18 crc kubenswrapper[4823]: I0320 09:21:18.808921 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rb9v" event={"ID":"4983947e-6154-4664-a5f0-9b5631076517","Type":"ContainerStarted","Data":"ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3"} Mar 20 09:21:18 crc kubenswrapper[4823]: I0320 09:21:18.842425 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2rb9v" podStartSLOduration=3.089474296 podStartE2EDuration="6.842398052s" podCreationTimestamp="2026-03-20 09:21:12 +0000 UTC" firstStartedPulling="2026-03-20 09:21:14.507481245 +0000 UTC m=+1441.193377619" lastFinishedPulling="2026-03-20 09:21:18.260405001 +0000 UTC m=+1444.946301375" observedRunningTime="2026-03-20 09:21:18.836062283 +0000 UTC m=+1445.521958697" watchObservedRunningTime="2026-03-20 09:21:18.842398052 +0000 UTC m=+1445.528294466" Mar 20 09:21:23 crc kubenswrapper[4823]: I0320 09:21:23.205462 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:23 crc kubenswrapper[4823]: I0320 09:21:23.206130 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:24 crc kubenswrapper[4823]: I0320 09:21:24.245209 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2rb9v" podUID="4983947e-6154-4664-a5f0-9b5631076517" containerName="registry-server" probeResult="failure" output=< Mar 20 09:21:24 crc kubenswrapper[4823]: timeout: failed to connect service ":50051" within 1s Mar 20 09:21:24 crc kubenswrapper[4823]: > Mar 20 09:21:33 crc kubenswrapper[4823]: I0320 09:21:33.265524 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:33 crc kubenswrapper[4823]: I0320 09:21:33.323404 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:33 crc kubenswrapper[4823]: I0320 09:21:33.503578 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2rb9v"] Mar 20 09:21:34 crc kubenswrapper[4823]: I0320 09:21:34.913368 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2rb9v" podUID="4983947e-6154-4664-a5f0-9b5631076517" containerName="registry-server" containerID="cri-o://ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3" gracePeriod=2 Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.329765 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.433411 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-catalog-content\") pod \"4983947e-6154-4664-a5f0-9b5631076517\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.433466 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-utilities\") pod \"4983947e-6154-4664-a5f0-9b5631076517\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.434536 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-utilities" (OuterVolumeSpecName: "utilities") pod "4983947e-6154-4664-a5f0-9b5631076517" (UID: "4983947e-6154-4664-a5f0-9b5631076517"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.434615 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jddg\" (UniqueName: \"kubernetes.io/projected/4983947e-6154-4664-a5f0-9b5631076517-kube-api-access-4jddg\") pod \"4983947e-6154-4664-a5f0-9b5631076517\" (UID: \"4983947e-6154-4664-a5f0-9b5631076517\") " Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.435997 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.440406 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4983947e-6154-4664-a5f0-9b5631076517-kube-api-access-4jddg" (OuterVolumeSpecName: "kube-api-access-4jddg") pod "4983947e-6154-4664-a5f0-9b5631076517" (UID: "4983947e-6154-4664-a5f0-9b5631076517"). InnerVolumeSpecName "kube-api-access-4jddg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.536912 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jddg\" (UniqueName: \"kubernetes.io/projected/4983947e-6154-4664-a5f0-9b5631076517-kube-api-access-4jddg\") on node \"crc\" DevicePath \"\"" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.563141 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4983947e-6154-4664-a5f0-9b5631076517" (UID: "4983947e-6154-4664-a5f0-9b5631076517"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.637595 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4983947e-6154-4664-a5f0-9b5631076517-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.920665 4823 generic.go:334] "Generic (PLEG): container finished" podID="4983947e-6154-4664-a5f0-9b5631076517" containerID="ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3" exitCode=0 Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.920707 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rb9v" event={"ID":"4983947e-6154-4664-a5f0-9b5631076517","Type":"ContainerDied","Data":"ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3"} Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.920731 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rb9v" event={"ID":"4983947e-6154-4664-a5f0-9b5631076517","Type":"ContainerDied","Data":"ccdc6ec5a38ea640b9e619f3f4c52d2a74f2aa78fca0e239dfe7ce0e9a25d0c7"} Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.920771 4823 scope.go:117] "RemoveContainer" containerID="ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.920887 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rb9v" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.938208 4823 scope.go:117] "RemoveContainer" containerID="2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.951516 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2rb9v"] Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.957505 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2rb9v"] Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.959975 4823 scope.go:117] "RemoveContainer" containerID="dc7206c2c104cd400cb12bb47a9f58bc6cb68291cf9e14dad33ff503a653061e" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.980722 4823 scope.go:117] "RemoveContainer" containerID="ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3" Mar 20 09:21:35 crc kubenswrapper[4823]: E0320 09:21:35.982926 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3\": container with ID starting with ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3 not found: ID does not exist" containerID="ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.982963 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3"} err="failed to get container status \"ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3\": rpc error: code = NotFound desc = could not find container \"ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3\": container with ID starting with ca45ee24f562117b1085b32956945f3892ff85bb5a20fc5528613509a3d02bb3 not found: ID does not exist" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.982990 4823 scope.go:117] "RemoveContainer" containerID="2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1" Mar 20 09:21:35 crc kubenswrapper[4823]: E0320 09:21:35.983386 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1\": container with ID starting with 2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1 not found: ID does not exist" containerID="2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.983435 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1"} err="failed to get container status \"2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1\": rpc error: code = NotFound desc = could not find container \"2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1\": container with ID starting with 2c3ea636044a830bfb92d95f2968cee5411641d8dc9806d62071d36280f80ef1 not found: ID does not exist" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.983467 4823 scope.go:117] "RemoveContainer" containerID="dc7206c2c104cd400cb12bb47a9f58bc6cb68291cf9e14dad33ff503a653061e" Mar 20 09:21:35 crc kubenswrapper[4823]: E0320 09:21:35.984061 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc7206c2c104cd400cb12bb47a9f58bc6cb68291cf9e14dad33ff503a653061e\": container with ID starting with dc7206c2c104cd400cb12bb47a9f58bc6cb68291cf9e14dad33ff503a653061e not found: ID does not exist" containerID="dc7206c2c104cd400cb12bb47a9f58bc6cb68291cf9e14dad33ff503a653061e" Mar 20 09:21:35 crc kubenswrapper[4823]: I0320 09:21:35.984091 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7206c2c104cd400cb12bb47a9f58bc6cb68291cf9e14dad33ff503a653061e"} err="failed to get container status \"dc7206c2c104cd400cb12bb47a9f58bc6cb68291cf9e14dad33ff503a653061e\": rpc error: code = NotFound desc = could not find container \"dc7206c2c104cd400cb12bb47a9f58bc6cb68291cf9e14dad33ff503a653061e\": container with ID starting with dc7206c2c104cd400cb12bb47a9f58bc6cb68291cf9e14dad33ff503a653061e not found: ID does not exist" Mar 20 09:21:36 crc kubenswrapper[4823]: I0320 09:21:36.104849 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4983947e-6154-4664-a5f0-9b5631076517" path="/var/lib/kubelet/pods/4983947e-6154-4664-a5f0-9b5631076517/volumes" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.155487 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566642-4n6qx"] Mar 20 09:22:00 crc kubenswrapper[4823]: E0320 09:22:00.158468 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4983947e-6154-4664-a5f0-9b5631076517" containerName="extract-utilities" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.158630 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="4983947e-6154-4664-a5f0-9b5631076517" containerName="extract-utilities" Mar 20 09:22:00 crc kubenswrapper[4823]: E0320 09:22:00.158826 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4983947e-6154-4664-a5f0-9b5631076517" containerName="extract-content" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.158980 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="4983947e-6154-4664-a5f0-9b5631076517" containerName="extract-content" Mar 20 09:22:00 crc kubenswrapper[4823]: E0320 09:22:00.159109 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4983947e-6154-4664-a5f0-9b5631076517" containerName="registry-server" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.159238 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="4983947e-6154-4664-a5f0-9b5631076517" containerName="registry-server" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.159658 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="4983947e-6154-4664-a5f0-9b5631076517" containerName="registry-server" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.160538 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566642-4n6qx" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.160931 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566642-4n6qx"] Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.197941 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.198124 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.198203 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.199732 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-466zb\" (UniqueName: \"kubernetes.io/projected/7b26860f-3bd2-4ab6-99f3-77fe0f0313eb-kube-api-access-466zb\") pod \"auto-csr-approver-29566642-4n6qx\" (UID: \"7b26860f-3bd2-4ab6-99f3-77fe0f0313eb\") " pod="openshift-infra/auto-csr-approver-29566642-4n6qx" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.301119 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-466zb\" (UniqueName: \"kubernetes.io/projected/7b26860f-3bd2-4ab6-99f3-77fe0f0313eb-kube-api-access-466zb\") pod \"auto-csr-approver-29566642-4n6qx\" (UID: \"7b26860f-3bd2-4ab6-99f3-77fe0f0313eb\") " pod="openshift-infra/auto-csr-approver-29566642-4n6qx" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.321348 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-466zb\" (UniqueName: \"kubernetes.io/projected/7b26860f-3bd2-4ab6-99f3-77fe0f0313eb-kube-api-access-466zb\") pod \"auto-csr-approver-29566642-4n6qx\" (UID: \"7b26860f-3bd2-4ab6-99f3-77fe0f0313eb\") " pod="openshift-infra/auto-csr-approver-29566642-4n6qx" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.525001 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566642-4n6qx" Mar 20 09:22:00 crc kubenswrapper[4823]: I0320 09:22:00.744349 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566642-4n6qx"] Mar 20 09:22:00 crc kubenswrapper[4823]: W0320 09:22:00.753590 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b26860f_3bd2_4ab6_99f3_77fe0f0313eb.slice/crio-7cdfcddc5ca13e0ca50ca1d991d20341d0952963cb2f81b2f12c5f484e9b19b6 WatchSource:0}: Error finding container 7cdfcddc5ca13e0ca50ca1d991d20341d0952963cb2f81b2f12c5f484e9b19b6: Status 404 returned error can't find the container with id 7cdfcddc5ca13e0ca50ca1d991d20341d0952963cb2f81b2f12c5f484e9b19b6 Mar 20 09:22:01 crc kubenswrapper[4823]: I0320 09:22:01.110794 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566642-4n6qx" event={"ID":"7b26860f-3bd2-4ab6-99f3-77fe0f0313eb","Type":"ContainerStarted","Data":"7cdfcddc5ca13e0ca50ca1d991d20341d0952963cb2f81b2f12c5f484e9b19b6"} Mar 20 09:22:02 crc kubenswrapper[4823]: I0320 09:22:02.119083 4823 generic.go:334] "Generic (PLEG): container finished" podID="7b26860f-3bd2-4ab6-99f3-77fe0f0313eb" containerID="39353325502cf8e868a07ec1ac521fa943a48d36078f0b38f9c88a1f4a0f2816" exitCode=0 Mar 20 09:22:02 crc kubenswrapper[4823]: I0320 09:22:02.119140 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566642-4n6qx" event={"ID":"7b26860f-3bd2-4ab6-99f3-77fe0f0313eb","Type":"ContainerDied","Data":"39353325502cf8e868a07ec1ac521fa943a48d36078f0b38f9c88a1f4a0f2816"} Mar 20 09:22:03 crc kubenswrapper[4823]: I0320 09:22:03.449092 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566642-4n6qx" Mar 20 09:22:03 crc kubenswrapper[4823]: I0320 09:22:03.647133 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-466zb\" (UniqueName: \"kubernetes.io/projected/7b26860f-3bd2-4ab6-99f3-77fe0f0313eb-kube-api-access-466zb\") pod \"7b26860f-3bd2-4ab6-99f3-77fe0f0313eb\" (UID: \"7b26860f-3bd2-4ab6-99f3-77fe0f0313eb\") " Mar 20 09:22:03 crc kubenswrapper[4823]: I0320 09:22:03.655420 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b26860f-3bd2-4ab6-99f3-77fe0f0313eb-kube-api-access-466zb" (OuterVolumeSpecName: "kube-api-access-466zb") pod "7b26860f-3bd2-4ab6-99f3-77fe0f0313eb" (UID: "7b26860f-3bd2-4ab6-99f3-77fe0f0313eb"). InnerVolumeSpecName "kube-api-access-466zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:22:03 crc kubenswrapper[4823]: I0320 09:22:03.748852 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-466zb\" (UniqueName: \"kubernetes.io/projected/7b26860f-3bd2-4ab6-99f3-77fe0f0313eb-kube-api-access-466zb\") on node \"crc\" DevicePath \"\"" Mar 20 09:22:04 crc kubenswrapper[4823]: I0320 09:22:04.138405 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566642-4n6qx" event={"ID":"7b26860f-3bd2-4ab6-99f3-77fe0f0313eb","Type":"ContainerDied","Data":"7cdfcddc5ca13e0ca50ca1d991d20341d0952963cb2f81b2f12c5f484e9b19b6"} Mar 20 09:22:04 crc kubenswrapper[4823]: I0320 09:22:04.138466 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566642-4n6qx" Mar 20 09:22:04 crc kubenswrapper[4823]: I0320 09:22:04.138468 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cdfcddc5ca13e0ca50ca1d991d20341d0952963cb2f81b2f12c5f484e9b19b6" Mar 20 09:22:04 crc kubenswrapper[4823]: I0320 09:22:04.520791 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566636-xtzck"] Mar 20 09:22:04 crc kubenswrapper[4823]: I0320 09:22:04.528423 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566636-xtzck"] Mar 20 09:22:06 crc kubenswrapper[4823]: I0320 09:22:06.111100 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cf48399-a918-4f54-89b2-0891d06d043b" path="/var/lib/kubelet/pods/5cf48399-a918-4f54-89b2-0891d06d043b/volumes" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.721935 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cd8pv"] Mar 20 09:22:19 crc kubenswrapper[4823]: E0320 09:22:19.723693 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b26860f-3bd2-4ab6-99f3-77fe0f0313eb" containerName="oc" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.723719 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b26860f-3bd2-4ab6-99f3-77fe0f0313eb" containerName="oc" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.724006 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b26860f-3bd2-4ab6-99f3-77fe0f0313eb" containerName="oc" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.727760 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.737924 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cd8pv"] Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.878766 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgnt6\" (UniqueName: \"kubernetes.io/projected/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-kube-api-access-lgnt6\") pod \"certified-operators-cd8pv\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.878937 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-catalog-content\") pod \"certified-operators-cd8pv\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.878971 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-utilities\") pod \"certified-operators-cd8pv\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.979670 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgnt6\" (UniqueName: \"kubernetes.io/projected/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-kube-api-access-lgnt6\") pod \"certified-operators-cd8pv\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.980047 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-catalog-content\") pod \"certified-operators-cd8pv\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.980159 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-utilities\") pod \"certified-operators-cd8pv\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.980606 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-catalog-content\") pod \"certified-operators-cd8pv\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:19 crc kubenswrapper[4823]: I0320 09:22:19.980803 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-utilities\") pod \"certified-operators-cd8pv\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:20 crc kubenswrapper[4823]: I0320 09:22:19.999089 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgnt6\" (UniqueName: \"kubernetes.io/projected/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-kube-api-access-lgnt6\") pod \"certified-operators-cd8pv\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:20 crc kubenswrapper[4823]: I0320 09:22:20.048757 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:20 crc kubenswrapper[4823]: I0320 09:22:20.535454 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cd8pv"] Mar 20 09:22:21 crc kubenswrapper[4823]: I0320 09:22:21.262539 4823 generic.go:334] "Generic (PLEG): container finished" podID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" containerID="b77d6710029366377f81e34d1d8ce531562fb501d07837ba801ba28f3f66fdb8" exitCode=0 Mar 20 09:22:21 crc kubenswrapper[4823]: I0320 09:22:21.262599 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd8pv" event={"ID":"77a35ed8-b18b-477e-a3e6-dc749a2b6caa","Type":"ContainerDied","Data":"b77d6710029366377f81e34d1d8ce531562fb501d07837ba801ba28f3f66fdb8"} Mar 20 09:22:21 crc kubenswrapper[4823]: I0320 09:22:21.262889 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd8pv" event={"ID":"77a35ed8-b18b-477e-a3e6-dc749a2b6caa","Type":"ContainerStarted","Data":"a75fff46a45d3fa3bc8fec5e55ed6571df9c7e3e5f6c683adfff767ff3f84fb9"} Mar 20 09:22:22 crc kubenswrapper[4823]: I0320 09:22:22.274106 4823 generic.go:334] "Generic (PLEG): container finished" podID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" containerID="8975d13d8dbbe7c0b239901ff54e9ded96e1237e7a2ca6cbf7b3dee29c7d8618" exitCode=0 Mar 20 09:22:22 crc kubenswrapper[4823]: I0320 09:22:22.274182 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd8pv" event={"ID":"77a35ed8-b18b-477e-a3e6-dc749a2b6caa","Type":"ContainerDied","Data":"8975d13d8dbbe7c0b239901ff54e9ded96e1237e7a2ca6cbf7b3dee29c7d8618"} Mar 20 09:22:23 crc kubenswrapper[4823]: I0320 09:22:23.288090 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd8pv" event={"ID":"77a35ed8-b18b-477e-a3e6-dc749a2b6caa","Type":"ContainerStarted","Data":"e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f"} Mar 20 09:22:23 crc kubenswrapper[4823]: I0320 09:22:23.306802 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cd8pv" podStartSLOduration=2.767912001 podStartE2EDuration="4.305969216s" podCreationTimestamp="2026-03-20 09:22:19 +0000 UTC" firstStartedPulling="2026-03-20 09:22:21.263874847 +0000 UTC m=+1507.949771221" lastFinishedPulling="2026-03-20 09:22:22.801932062 +0000 UTC m=+1509.487828436" observedRunningTime="2026-03-20 09:22:23.304453997 +0000 UTC m=+1509.990350381" watchObservedRunningTime="2026-03-20 09:22:23.305969216 +0000 UTC m=+1509.991865600" Mar 20 09:22:30 crc kubenswrapper[4823]: I0320 09:22:30.049899 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:30 crc kubenswrapper[4823]: I0320 09:22:30.050627 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:30 crc kubenswrapper[4823]: I0320 09:22:30.091479 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:30 crc kubenswrapper[4823]: I0320 09:22:30.380095 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:30 crc kubenswrapper[4823]: I0320 09:22:30.419781 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cd8pv"] Mar 20 09:22:32 crc kubenswrapper[4823]: I0320 09:22:32.351645 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cd8pv" podUID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" containerName="registry-server" containerID="cri-o://e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f" gracePeriod=2 Mar 20 09:22:32 crc kubenswrapper[4823]: I0320 09:22:32.751628 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:32 crc kubenswrapper[4823]: I0320 09:22:32.905270 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgnt6\" (UniqueName: \"kubernetes.io/projected/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-kube-api-access-lgnt6\") pod \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " Mar 20 09:22:32 crc kubenswrapper[4823]: I0320 09:22:32.905416 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-utilities\") pod \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " Mar 20 09:22:32 crc kubenswrapper[4823]: I0320 09:22:32.905479 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-catalog-content\") pod \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\" (UID: \"77a35ed8-b18b-477e-a3e6-dc749a2b6caa\") " Mar 20 09:22:32 crc kubenswrapper[4823]: I0320 09:22:32.906206 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-utilities" (OuterVolumeSpecName: "utilities") pod "77a35ed8-b18b-477e-a3e6-dc749a2b6caa" (UID: "77a35ed8-b18b-477e-a3e6-dc749a2b6caa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:22:32 crc kubenswrapper[4823]: I0320 09:22:32.911044 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-kube-api-access-lgnt6" (OuterVolumeSpecName: "kube-api-access-lgnt6") pod "77a35ed8-b18b-477e-a3e6-dc749a2b6caa" (UID: "77a35ed8-b18b-477e-a3e6-dc749a2b6caa"). InnerVolumeSpecName "kube-api-access-lgnt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.007476 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgnt6\" (UniqueName: \"kubernetes.io/projected/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-kube-api-access-lgnt6\") on node \"crc\" DevicePath \"\"" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.007517 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.360885 4823 generic.go:334] "Generic (PLEG): container finished" podID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" containerID="e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f" exitCode=0 Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.360926 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd8pv" event={"ID":"77a35ed8-b18b-477e-a3e6-dc749a2b6caa","Type":"ContainerDied","Data":"e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f"} Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.360975 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd8pv" event={"ID":"77a35ed8-b18b-477e-a3e6-dc749a2b6caa","Type":"ContainerDied","Data":"a75fff46a45d3fa3bc8fec5e55ed6571df9c7e3e5f6c683adfff767ff3f84fb9"} Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.360994 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd8pv" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.361014 4823 scope.go:117] "RemoveContainer" containerID="e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.381469 4823 scope.go:117] "RemoveContainer" containerID="8975d13d8dbbe7c0b239901ff54e9ded96e1237e7a2ca6cbf7b3dee29c7d8618" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.398100 4823 scope.go:117] "RemoveContainer" containerID="b77d6710029366377f81e34d1d8ce531562fb501d07837ba801ba28f3f66fdb8" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.431030 4823 scope.go:117] "RemoveContainer" containerID="e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f" Mar 20 09:22:33 crc kubenswrapper[4823]: E0320 09:22:33.431396 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f\": container with ID starting with e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f not found: ID does not exist" containerID="e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.431423 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f"} err="failed to get container status \"e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f\": rpc error: code = NotFound desc = could not find container \"e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f\": container with ID starting with e328c08a56f3872b94aee46502cf835c518d42a72d2204ba0c72aae05820f53f not found: ID does not exist" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.431441 4823 scope.go:117] "RemoveContainer" containerID="8975d13d8dbbe7c0b239901ff54e9ded96e1237e7a2ca6cbf7b3dee29c7d8618" Mar 20 09:22:33 crc kubenswrapper[4823]: E0320 09:22:33.431721 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8975d13d8dbbe7c0b239901ff54e9ded96e1237e7a2ca6cbf7b3dee29c7d8618\": container with ID starting with 8975d13d8dbbe7c0b239901ff54e9ded96e1237e7a2ca6cbf7b3dee29c7d8618 not found: ID does not exist" containerID="8975d13d8dbbe7c0b239901ff54e9ded96e1237e7a2ca6cbf7b3dee29c7d8618" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.431753 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8975d13d8dbbe7c0b239901ff54e9ded96e1237e7a2ca6cbf7b3dee29c7d8618"} err="failed to get container status \"8975d13d8dbbe7c0b239901ff54e9ded96e1237e7a2ca6cbf7b3dee29c7d8618\": rpc error: code = NotFound desc = could not find container \"8975d13d8dbbe7c0b239901ff54e9ded96e1237e7a2ca6cbf7b3dee29c7d8618\": container with ID starting with 8975d13d8dbbe7c0b239901ff54e9ded96e1237e7a2ca6cbf7b3dee29c7d8618 not found: ID does not exist" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.431765 4823 scope.go:117] "RemoveContainer" containerID="b77d6710029366377f81e34d1d8ce531562fb501d07837ba801ba28f3f66fdb8" Mar 20 09:22:33 crc kubenswrapper[4823]: E0320 09:22:33.431987 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b77d6710029366377f81e34d1d8ce531562fb501d07837ba801ba28f3f66fdb8\": container with ID starting with b77d6710029366377f81e34d1d8ce531562fb501d07837ba801ba28f3f66fdb8 not found: ID does not exist" containerID="b77d6710029366377f81e34d1d8ce531562fb501d07837ba801ba28f3f66fdb8" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.432007 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b77d6710029366377f81e34d1d8ce531562fb501d07837ba801ba28f3f66fdb8"} err="failed to get container status \"b77d6710029366377f81e34d1d8ce531562fb501d07837ba801ba28f3f66fdb8\": rpc error: code = NotFound desc = could not find container \"b77d6710029366377f81e34d1d8ce531562fb501d07837ba801ba28f3f66fdb8\": container with ID starting with b77d6710029366377f81e34d1d8ce531562fb501d07837ba801ba28f3f66fdb8 not found: ID does not exist" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.443728 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77a35ed8-b18b-477e-a3e6-dc749a2b6caa" (UID: "77a35ed8-b18b-477e-a3e6-dc749a2b6caa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.514536 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77a35ed8-b18b-477e-a3e6-dc749a2b6caa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.706359 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cd8pv"] Mar 20 09:22:33 crc kubenswrapper[4823]: I0320 09:22:33.711671 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cd8pv"] Mar 20 09:22:34 crc kubenswrapper[4823]: I0320 09:22:34.107533 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" path="/var/lib/kubelet/pods/77a35ed8-b18b-477e-a3e6-dc749a2b6caa/volumes" Mar 20 09:22:38 crc kubenswrapper[4823]: I0320 09:22:38.842827 4823 scope.go:117] "RemoveContainer" containerID="0b249db9d3205e23cb180145a20b04e976969d5ccb423d493b6bed25f85cea37" Mar 20 09:22:39 crc kubenswrapper[4823]: I0320 09:22:39.430272 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:22:39 crc kubenswrapper[4823]: I0320 09:22:39.430322 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:23:09 crc kubenswrapper[4823]: I0320 09:23:09.430928 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:23:09 crc kubenswrapper[4823]: I0320 09:23:09.431605 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:23:39 crc kubenswrapper[4823]: I0320 09:23:39.430471 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:23:39 crc kubenswrapper[4823]: I0320 09:23:39.431152 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:23:39 crc kubenswrapper[4823]: I0320 09:23:39.431201 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:23:39 crc kubenswrapper[4823]: I0320 09:23:39.431869 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:23:39 crc kubenswrapper[4823]: I0320 09:23:39.431919 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" gracePeriod=600 Mar 20 09:23:39 crc kubenswrapper[4823]: E0320 09:23:39.554592 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:23:39 crc kubenswrapper[4823]: I0320 09:23:39.878870 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" exitCode=0 Mar 20 09:23:39 crc kubenswrapper[4823]: I0320 09:23:39.878925 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8"} Mar 20 09:23:39 crc kubenswrapper[4823]: I0320 09:23:39.878971 4823 scope.go:117] "RemoveContainer" containerID="78975c9ca0024f8f8c44d3ccb13c9eb4809cc513d430445cccc664f308a39c85" Mar 20 09:23:39 crc kubenswrapper[4823]: I0320 09:23:39.879563 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:23:39 crc kubenswrapper[4823]: E0320 09:23:39.879860 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:23:53 crc kubenswrapper[4823]: I0320 09:23:53.097000 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:23:53 crc kubenswrapper[4823]: E0320 09:23:53.098106 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.146830 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566644-xsgdd"] Mar 20 09:24:00 crc kubenswrapper[4823]: E0320 09:24:00.147827 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" containerName="extract-content" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.147844 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" containerName="extract-content" Mar 20 09:24:00 crc kubenswrapper[4823]: E0320 09:24:00.147857 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" containerName="registry-server" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.147864 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" containerName="registry-server" Mar 20 09:24:00 crc kubenswrapper[4823]: E0320 09:24:00.147882 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" containerName="extract-utilities" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.147890 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" containerName="extract-utilities" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.148017 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="77a35ed8-b18b-477e-a3e6-dc749a2b6caa" containerName="registry-server" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.148602 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566644-xsgdd" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.151694 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.151908 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.152163 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.162450 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566644-xsgdd"] Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.288486 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8k5k\" (UniqueName: \"kubernetes.io/projected/2aa6bd30-bf9b-42a4-956e-ea22d0c15335-kube-api-access-q8k5k\") pod \"auto-csr-approver-29566644-xsgdd\" (UID: \"2aa6bd30-bf9b-42a4-956e-ea22d0c15335\") " pod="openshift-infra/auto-csr-approver-29566644-xsgdd" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.389605 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8k5k\" (UniqueName: \"kubernetes.io/projected/2aa6bd30-bf9b-42a4-956e-ea22d0c15335-kube-api-access-q8k5k\") pod \"auto-csr-approver-29566644-xsgdd\" (UID: \"2aa6bd30-bf9b-42a4-956e-ea22d0c15335\") " pod="openshift-infra/auto-csr-approver-29566644-xsgdd" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.413490 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8k5k\" (UniqueName: \"kubernetes.io/projected/2aa6bd30-bf9b-42a4-956e-ea22d0c15335-kube-api-access-q8k5k\") pod \"auto-csr-approver-29566644-xsgdd\" (UID: \"2aa6bd30-bf9b-42a4-956e-ea22d0c15335\") " pod="openshift-infra/auto-csr-approver-29566644-xsgdd" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.469464 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566644-xsgdd" Mar 20 09:24:00 crc kubenswrapper[4823]: I0320 09:24:00.928318 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566644-xsgdd"] Mar 20 09:24:00 crc kubenswrapper[4823]: W0320 09:24:00.942997 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2aa6bd30_bf9b_42a4_956e_ea22d0c15335.slice/crio-d48277d68719e922a5d05ca6429fe0de3632c51ac7c48558e82549cd794a1614 WatchSource:0}: Error finding container d48277d68719e922a5d05ca6429fe0de3632c51ac7c48558e82549cd794a1614: Status 404 returned error can't find the container with id d48277d68719e922a5d05ca6429fe0de3632c51ac7c48558e82549cd794a1614 Mar 20 09:24:01 crc kubenswrapper[4823]: I0320 09:24:01.055172 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566644-xsgdd" event={"ID":"2aa6bd30-bf9b-42a4-956e-ea22d0c15335","Type":"ContainerStarted","Data":"d48277d68719e922a5d05ca6429fe0de3632c51ac7c48558e82549cd794a1614"} Mar 20 09:24:03 crc kubenswrapper[4823]: I0320 09:24:03.073091 4823 generic.go:334] "Generic (PLEG): container finished" podID="2aa6bd30-bf9b-42a4-956e-ea22d0c15335" containerID="65940099fccc9b3ecfacfc06580c6355bdd24d7f90213ce10826b151aa46c696" exitCode=0 Mar 20 09:24:03 crc kubenswrapper[4823]: I0320 09:24:03.073171 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566644-xsgdd" event={"ID":"2aa6bd30-bf9b-42a4-956e-ea22d0c15335","Type":"ContainerDied","Data":"65940099fccc9b3ecfacfc06580c6355bdd24d7f90213ce10826b151aa46c696"} Mar 20 09:24:04 crc kubenswrapper[4823]: I0320 09:24:04.378965 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566644-xsgdd" Mar 20 09:24:04 crc kubenswrapper[4823]: I0320 09:24:04.443708 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8k5k\" (UniqueName: \"kubernetes.io/projected/2aa6bd30-bf9b-42a4-956e-ea22d0c15335-kube-api-access-q8k5k\") pod \"2aa6bd30-bf9b-42a4-956e-ea22d0c15335\" (UID: \"2aa6bd30-bf9b-42a4-956e-ea22d0c15335\") " Mar 20 09:24:04 crc kubenswrapper[4823]: I0320 09:24:04.451235 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aa6bd30-bf9b-42a4-956e-ea22d0c15335-kube-api-access-q8k5k" (OuterVolumeSpecName: "kube-api-access-q8k5k") pod "2aa6bd30-bf9b-42a4-956e-ea22d0c15335" (UID: "2aa6bd30-bf9b-42a4-956e-ea22d0c15335"). InnerVolumeSpecName "kube-api-access-q8k5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:24:04 crc kubenswrapper[4823]: I0320 09:24:04.545350 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8k5k\" (UniqueName: \"kubernetes.io/projected/2aa6bd30-bf9b-42a4-956e-ea22d0c15335-kube-api-access-q8k5k\") on node \"crc\" DevicePath \"\"" Mar 20 09:24:05 crc kubenswrapper[4823]: I0320 09:24:05.092610 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566644-xsgdd" event={"ID":"2aa6bd30-bf9b-42a4-956e-ea22d0c15335","Type":"ContainerDied","Data":"d48277d68719e922a5d05ca6429fe0de3632c51ac7c48558e82549cd794a1614"} Mar 20 09:24:05 crc kubenswrapper[4823]: I0320 09:24:05.092972 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d48277d68719e922a5d05ca6429fe0de3632c51ac7c48558e82549cd794a1614" Mar 20 09:24:05 crc kubenswrapper[4823]: I0320 09:24:05.092657 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566644-xsgdd" Mar 20 09:24:05 crc kubenswrapper[4823]: I0320 09:24:05.454929 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566638-vf9vh"] Mar 20 09:24:05 crc kubenswrapper[4823]: I0320 09:24:05.460028 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566638-vf9vh"] Mar 20 09:24:06 crc kubenswrapper[4823]: I0320 09:24:06.109888 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="834573f2-b90a-48f0-97ab-f5cff6141a90" path="/var/lib/kubelet/pods/834573f2-b90a-48f0-97ab-f5cff6141a90/volumes" Mar 20 09:24:06 crc kubenswrapper[4823]: I0320 09:24:06.112130 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:24:06 crc kubenswrapper[4823]: E0320 09:24:06.112603 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:24:17 crc kubenswrapper[4823]: I0320 09:24:17.097602 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:24:17 crc kubenswrapper[4823]: E0320 09:24:17.098892 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:24:28 crc kubenswrapper[4823]: I0320 09:24:28.096838 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:24:28 crc kubenswrapper[4823]: E0320 09:24:28.097703 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:24:38 crc kubenswrapper[4823]: I0320 09:24:38.934963 4823 scope.go:117] "RemoveContainer" containerID="79014211ee87220c7250d3e5d4753caf8f54863a176aeaf51bedbe4a5ec4fc6a" Mar 20 09:24:43 crc kubenswrapper[4823]: I0320 09:24:43.097563 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:24:43 crc kubenswrapper[4823]: E0320 09:24:43.098200 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:24:54 crc kubenswrapper[4823]: I0320 09:24:54.102442 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:24:54 crc kubenswrapper[4823]: E0320 09:24:54.103731 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:25:08 crc kubenswrapper[4823]: I0320 09:25:08.097364 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:25:08 crc kubenswrapper[4823]: E0320 09:25:08.098233 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:25:19 crc kubenswrapper[4823]: I0320 09:25:19.097096 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:25:19 crc kubenswrapper[4823]: E0320 09:25:19.099202 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:25:32 crc kubenswrapper[4823]: I0320 09:25:32.097663 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:25:32 crc kubenswrapper[4823]: E0320 09:25:32.099199 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.004130 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kpb2h"] Mar 20 09:25:33 crc kubenswrapper[4823]: E0320 09:25:33.004924 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa6bd30-bf9b-42a4-956e-ea22d0c15335" containerName="oc" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.004949 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa6bd30-bf9b-42a4-956e-ea22d0c15335" containerName="oc" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.005106 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aa6bd30-bf9b-42a4-956e-ea22d0c15335" containerName="oc" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.006487 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.012073 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpb2h"] Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.041245 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2x84\" (UniqueName: \"kubernetes.io/projected/bdbf89d1-dd38-4383-a9ae-94c32689740d-kube-api-access-s2x84\") pod \"redhat-marketplace-kpb2h\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.041304 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-utilities\") pod \"redhat-marketplace-kpb2h\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.041553 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-catalog-content\") pod \"redhat-marketplace-kpb2h\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.142658 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2x84\" (UniqueName: \"kubernetes.io/projected/bdbf89d1-dd38-4383-a9ae-94c32689740d-kube-api-access-s2x84\") pod \"redhat-marketplace-kpb2h\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.142792 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-utilities\") pod \"redhat-marketplace-kpb2h\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.142882 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-catalog-content\") pod \"redhat-marketplace-kpb2h\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.143619 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-utilities\") pod \"redhat-marketplace-kpb2h\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.144238 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-catalog-content\") pod \"redhat-marketplace-kpb2h\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.163920 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2x84\" (UniqueName: \"kubernetes.io/projected/bdbf89d1-dd38-4383-a9ae-94c32689740d-kube-api-access-s2x84\") pod \"redhat-marketplace-kpb2h\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.357370 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:33 crc kubenswrapper[4823]: I0320 09:25:33.779762 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpb2h"] Mar 20 09:25:34 crc kubenswrapper[4823]: I0320 09:25:34.785102 4823 generic.go:334] "Generic (PLEG): container finished" podID="bdbf89d1-dd38-4383-a9ae-94c32689740d" containerID="4d064c0f28fb2003393ec271c35f3b98799da8f87006419ddbec1b07413267aa" exitCode=0 Mar 20 09:25:34 crc kubenswrapper[4823]: I0320 09:25:34.785434 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpb2h" event={"ID":"bdbf89d1-dd38-4383-a9ae-94c32689740d","Type":"ContainerDied","Data":"4d064c0f28fb2003393ec271c35f3b98799da8f87006419ddbec1b07413267aa"} Mar 20 09:25:34 crc kubenswrapper[4823]: I0320 09:25:34.785463 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpb2h" event={"ID":"bdbf89d1-dd38-4383-a9ae-94c32689740d","Type":"ContainerStarted","Data":"cd482e44d7c084782691139fdd0c40267289ba211b6514c071cbef0c2632e3a5"} Mar 20 09:25:35 crc kubenswrapper[4823]: I0320 09:25:35.794200 4823 generic.go:334] "Generic (PLEG): container finished" podID="bdbf89d1-dd38-4383-a9ae-94c32689740d" containerID="696a8bd67330bc6928f86a69ae79383f4f48b0776924a5b5d0d672596e3b2df5" exitCode=0 Mar 20 09:25:35 crc kubenswrapper[4823]: I0320 09:25:35.794904 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpb2h" event={"ID":"bdbf89d1-dd38-4383-a9ae-94c32689740d","Type":"ContainerDied","Data":"696a8bd67330bc6928f86a69ae79383f4f48b0776924a5b5d0d672596e3b2df5"} Mar 20 09:25:36 crc kubenswrapper[4823]: I0320 09:25:36.804954 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpb2h" event={"ID":"bdbf89d1-dd38-4383-a9ae-94c32689740d","Type":"ContainerStarted","Data":"6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b"} Mar 20 09:25:36 crc kubenswrapper[4823]: I0320 09:25:36.834545 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kpb2h" podStartSLOduration=3.363148575 podStartE2EDuration="4.834522371s" podCreationTimestamp="2026-03-20 09:25:32 +0000 UTC" firstStartedPulling="2026-03-20 09:25:34.787176377 +0000 UTC m=+1701.473072751" lastFinishedPulling="2026-03-20 09:25:36.258550173 +0000 UTC m=+1702.944446547" observedRunningTime="2026-03-20 09:25:36.830437532 +0000 UTC m=+1703.516333946" watchObservedRunningTime="2026-03-20 09:25:36.834522371 +0000 UTC m=+1703.520418755" Mar 20 09:25:43 crc kubenswrapper[4823]: I0320 09:25:43.357496 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:43 crc kubenswrapper[4823]: I0320 09:25:43.358278 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:43 crc kubenswrapper[4823]: I0320 09:25:43.428046 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:43 crc kubenswrapper[4823]: I0320 09:25:43.911561 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:43 crc kubenswrapper[4823]: I0320 09:25:43.968830 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpb2h"] Mar 20 09:25:45 crc kubenswrapper[4823]: I0320 09:25:45.097050 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:25:45 crc kubenswrapper[4823]: E0320 09:25:45.097544 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:25:45 crc kubenswrapper[4823]: I0320 09:25:45.880344 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kpb2h" podUID="bdbf89d1-dd38-4383-a9ae-94c32689740d" containerName="registry-server" containerID="cri-o://6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b" gracePeriod=2 Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.345221 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.447209 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2x84\" (UniqueName: \"kubernetes.io/projected/bdbf89d1-dd38-4383-a9ae-94c32689740d-kube-api-access-s2x84\") pod \"bdbf89d1-dd38-4383-a9ae-94c32689740d\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.448292 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-catalog-content\") pod \"bdbf89d1-dd38-4383-a9ae-94c32689740d\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.449323 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-utilities\") pod \"bdbf89d1-dd38-4383-a9ae-94c32689740d\" (UID: \"bdbf89d1-dd38-4383-a9ae-94c32689740d\") " Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.450113 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-utilities" (OuterVolumeSpecName: "utilities") pod "bdbf89d1-dd38-4383-a9ae-94c32689740d" (UID: "bdbf89d1-dd38-4383-a9ae-94c32689740d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.454982 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdbf89d1-dd38-4383-a9ae-94c32689740d-kube-api-access-s2x84" (OuterVolumeSpecName: "kube-api-access-s2x84") pod "bdbf89d1-dd38-4383-a9ae-94c32689740d" (UID: "bdbf89d1-dd38-4383-a9ae-94c32689740d"). InnerVolumeSpecName "kube-api-access-s2x84". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.482008 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bdbf89d1-dd38-4383-a9ae-94c32689740d" (UID: "bdbf89d1-dd38-4383-a9ae-94c32689740d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.551487 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2x84\" (UniqueName: \"kubernetes.io/projected/bdbf89d1-dd38-4383-a9ae-94c32689740d-kube-api-access-s2x84\") on node \"crc\" DevicePath \"\"" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.551536 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.551555 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbf89d1-dd38-4383-a9ae-94c32689740d-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.893976 4823 generic.go:334] "Generic (PLEG): container finished" podID="bdbf89d1-dd38-4383-a9ae-94c32689740d" containerID="6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b" exitCode=0 Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.894034 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpb2h" event={"ID":"bdbf89d1-dd38-4383-a9ae-94c32689740d","Type":"ContainerDied","Data":"6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b"} Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.894084 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpb2h" event={"ID":"bdbf89d1-dd38-4383-a9ae-94c32689740d","Type":"ContainerDied","Data":"cd482e44d7c084782691139fdd0c40267289ba211b6514c071cbef0c2632e3a5"} Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.894107 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpb2h" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.894116 4823 scope.go:117] "RemoveContainer" containerID="6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.919376 4823 scope.go:117] "RemoveContainer" containerID="696a8bd67330bc6928f86a69ae79383f4f48b0776924a5b5d0d672596e3b2df5" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.943249 4823 scope.go:117] "RemoveContainer" containerID="4d064c0f28fb2003393ec271c35f3b98799da8f87006419ddbec1b07413267aa" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.944202 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpb2h"] Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.952062 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpb2h"] Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.989396 4823 scope.go:117] "RemoveContainer" containerID="6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b" Mar 20 09:25:46 crc kubenswrapper[4823]: E0320 09:25:46.989980 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b\": container with ID starting with 6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b not found: ID does not exist" containerID="6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.990033 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b"} err="failed to get container status \"6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b\": rpc error: code = NotFound desc = could not find container \"6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b\": container with ID starting with 6a151db4fe8f84ae79922d213a7e2d3e2e7494133c08a1e9440be11be682404b not found: ID does not exist" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.990071 4823 scope.go:117] "RemoveContainer" containerID="696a8bd67330bc6928f86a69ae79383f4f48b0776924a5b5d0d672596e3b2df5" Mar 20 09:25:46 crc kubenswrapper[4823]: E0320 09:25:46.990611 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"696a8bd67330bc6928f86a69ae79383f4f48b0776924a5b5d0d672596e3b2df5\": container with ID starting with 696a8bd67330bc6928f86a69ae79383f4f48b0776924a5b5d0d672596e3b2df5 not found: ID does not exist" containerID="696a8bd67330bc6928f86a69ae79383f4f48b0776924a5b5d0d672596e3b2df5" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.990648 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"696a8bd67330bc6928f86a69ae79383f4f48b0776924a5b5d0d672596e3b2df5"} err="failed to get container status \"696a8bd67330bc6928f86a69ae79383f4f48b0776924a5b5d0d672596e3b2df5\": rpc error: code = NotFound desc = could not find container \"696a8bd67330bc6928f86a69ae79383f4f48b0776924a5b5d0d672596e3b2df5\": container with ID starting with 696a8bd67330bc6928f86a69ae79383f4f48b0776924a5b5d0d672596e3b2df5 not found: ID does not exist" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.990673 4823 scope.go:117] "RemoveContainer" containerID="4d064c0f28fb2003393ec271c35f3b98799da8f87006419ddbec1b07413267aa" Mar 20 09:25:46 crc kubenswrapper[4823]: E0320 09:25:46.991023 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d064c0f28fb2003393ec271c35f3b98799da8f87006419ddbec1b07413267aa\": container with ID starting with 4d064c0f28fb2003393ec271c35f3b98799da8f87006419ddbec1b07413267aa not found: ID does not exist" containerID="4d064c0f28fb2003393ec271c35f3b98799da8f87006419ddbec1b07413267aa" Mar 20 09:25:46 crc kubenswrapper[4823]: I0320 09:25:46.991052 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d064c0f28fb2003393ec271c35f3b98799da8f87006419ddbec1b07413267aa"} err="failed to get container status \"4d064c0f28fb2003393ec271c35f3b98799da8f87006419ddbec1b07413267aa\": rpc error: code = NotFound desc = could not find container \"4d064c0f28fb2003393ec271c35f3b98799da8f87006419ddbec1b07413267aa\": container with ID starting with 4d064c0f28fb2003393ec271c35f3b98799da8f87006419ddbec1b07413267aa not found: ID does not exist" Mar 20 09:25:48 crc kubenswrapper[4823]: I0320 09:25:48.112577 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdbf89d1-dd38-4383-a9ae-94c32689740d" path="/var/lib/kubelet/pods/bdbf89d1-dd38-4383-a9ae-94c32689740d/volumes" Mar 20 09:25:58 crc kubenswrapper[4823]: I0320 09:25:58.097800 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:25:58 crc kubenswrapper[4823]: E0320 09:25:58.098761 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.147562 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566646-gx29f"] Mar 20 09:26:00 crc kubenswrapper[4823]: E0320 09:26:00.150049 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdbf89d1-dd38-4383-a9ae-94c32689740d" containerName="extract-utilities" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.150219 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdbf89d1-dd38-4383-a9ae-94c32689740d" containerName="extract-utilities" Mar 20 09:26:00 crc kubenswrapper[4823]: E0320 09:26:00.150397 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdbf89d1-dd38-4383-a9ae-94c32689740d" containerName="registry-server" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.150518 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdbf89d1-dd38-4383-a9ae-94c32689740d" containerName="registry-server" Mar 20 09:26:00 crc kubenswrapper[4823]: E0320 09:26:00.150666 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdbf89d1-dd38-4383-a9ae-94c32689740d" containerName="extract-content" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.150855 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdbf89d1-dd38-4383-a9ae-94c32689740d" containerName="extract-content" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.151223 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdbf89d1-dd38-4383-a9ae-94c32689740d" containerName="registry-server" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.152123 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566646-gx29f" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.156462 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.156711 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.157950 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.158685 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566646-gx29f"] Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.257997 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxqh4\" (UniqueName: \"kubernetes.io/projected/1591a99f-9fc4-4892-9637-298af3ea2e35-kube-api-access-pxqh4\") pod \"auto-csr-approver-29566646-gx29f\" (UID: \"1591a99f-9fc4-4892-9637-298af3ea2e35\") " pod="openshift-infra/auto-csr-approver-29566646-gx29f" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.359375 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxqh4\" (UniqueName: \"kubernetes.io/projected/1591a99f-9fc4-4892-9637-298af3ea2e35-kube-api-access-pxqh4\") pod \"auto-csr-approver-29566646-gx29f\" (UID: \"1591a99f-9fc4-4892-9637-298af3ea2e35\") " pod="openshift-infra/auto-csr-approver-29566646-gx29f" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.393541 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxqh4\" (UniqueName: \"kubernetes.io/projected/1591a99f-9fc4-4892-9637-298af3ea2e35-kube-api-access-pxqh4\") pod \"auto-csr-approver-29566646-gx29f\" (UID: \"1591a99f-9fc4-4892-9637-298af3ea2e35\") " pod="openshift-infra/auto-csr-approver-29566646-gx29f" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.487200 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566646-gx29f" Mar 20 09:26:00 crc kubenswrapper[4823]: I0320 09:26:00.940541 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566646-gx29f"] Mar 20 09:26:00 crc kubenswrapper[4823]: W0320 09:26:00.952439 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1591a99f_9fc4_4892_9637_298af3ea2e35.slice/crio-126d57b3d2c0b9e6b18866141ba40ad9e690aff77b285b7b28eeb006e0e1ab65 WatchSource:0}: Error finding container 126d57b3d2c0b9e6b18866141ba40ad9e690aff77b285b7b28eeb006e0e1ab65: Status 404 returned error can't find the container with id 126d57b3d2c0b9e6b18866141ba40ad9e690aff77b285b7b28eeb006e0e1ab65 Mar 20 09:26:01 crc kubenswrapper[4823]: I0320 09:26:01.013267 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566646-gx29f" event={"ID":"1591a99f-9fc4-4892-9637-298af3ea2e35","Type":"ContainerStarted","Data":"126d57b3d2c0b9e6b18866141ba40ad9e690aff77b285b7b28eeb006e0e1ab65"} Mar 20 09:26:03 crc kubenswrapper[4823]: I0320 09:26:03.035848 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566646-gx29f" event={"ID":"1591a99f-9fc4-4892-9637-298af3ea2e35","Type":"ContainerDied","Data":"e204714638a805ddb16b5444e92ae4cacd5c5dc8448c5a3beaa809db5648afee"} Mar 20 09:26:03 crc kubenswrapper[4823]: I0320 09:26:03.036021 4823 generic.go:334] "Generic (PLEG): container finished" podID="1591a99f-9fc4-4892-9637-298af3ea2e35" containerID="e204714638a805ddb16b5444e92ae4cacd5c5dc8448c5a3beaa809db5648afee" exitCode=0 Mar 20 09:26:04 crc kubenswrapper[4823]: I0320 09:26:04.306206 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566646-gx29f" Mar 20 09:26:04 crc kubenswrapper[4823]: I0320 09:26:04.415370 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxqh4\" (UniqueName: \"kubernetes.io/projected/1591a99f-9fc4-4892-9637-298af3ea2e35-kube-api-access-pxqh4\") pod \"1591a99f-9fc4-4892-9637-298af3ea2e35\" (UID: \"1591a99f-9fc4-4892-9637-298af3ea2e35\") " Mar 20 09:26:04 crc kubenswrapper[4823]: I0320 09:26:04.420868 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1591a99f-9fc4-4892-9637-298af3ea2e35-kube-api-access-pxqh4" (OuterVolumeSpecName: "kube-api-access-pxqh4") pod "1591a99f-9fc4-4892-9637-298af3ea2e35" (UID: "1591a99f-9fc4-4892-9637-298af3ea2e35"). InnerVolumeSpecName "kube-api-access-pxqh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:26:04 crc kubenswrapper[4823]: I0320 09:26:04.517773 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxqh4\" (UniqueName: \"kubernetes.io/projected/1591a99f-9fc4-4892-9637-298af3ea2e35-kube-api-access-pxqh4\") on node \"crc\" DevicePath \"\"" Mar 20 09:26:05 crc kubenswrapper[4823]: I0320 09:26:05.061047 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566646-gx29f" event={"ID":"1591a99f-9fc4-4892-9637-298af3ea2e35","Type":"ContainerDied","Data":"126d57b3d2c0b9e6b18866141ba40ad9e690aff77b285b7b28eeb006e0e1ab65"} Mar 20 09:26:05 crc kubenswrapper[4823]: I0320 09:26:05.061089 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566646-gx29f" Mar 20 09:26:05 crc kubenswrapper[4823]: I0320 09:26:05.061102 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="126d57b3d2c0b9e6b18866141ba40ad9e690aff77b285b7b28eeb006e0e1ab65" Mar 20 09:26:05 crc kubenswrapper[4823]: I0320 09:26:05.378654 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566640-5phwr"] Mar 20 09:26:05 crc kubenswrapper[4823]: I0320 09:26:05.383541 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566640-5phwr"] Mar 20 09:26:06 crc kubenswrapper[4823]: I0320 09:26:06.109967 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fac8eee5-c573-4911-b56b-f4804a790292" path="/var/lib/kubelet/pods/fac8eee5-c573-4911-b56b-f4804a790292/volumes" Mar 20 09:26:13 crc kubenswrapper[4823]: I0320 09:26:13.097189 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:26:13 crc kubenswrapper[4823]: E0320 09:26:13.098244 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:26:26 crc kubenswrapper[4823]: I0320 09:26:26.097574 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:26:26 crc kubenswrapper[4823]: E0320 09:26:26.098899 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:26:38 crc kubenswrapper[4823]: I0320 09:26:38.097435 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:26:38 crc kubenswrapper[4823]: E0320 09:26:38.098515 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:26:39 crc kubenswrapper[4823]: I0320 09:26:39.027036 4823 scope.go:117] "RemoveContainer" containerID="e77a907b22b846213e43d0cd198abd38f417f42e49737282b2d9e0a187a5551c" Mar 20 09:26:52 crc kubenswrapper[4823]: I0320 09:26:52.105003 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:26:52 crc kubenswrapper[4823]: E0320 09:26:52.105965 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:27:03 crc kubenswrapper[4823]: I0320 09:27:03.097927 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:27:03 crc kubenswrapper[4823]: E0320 09:27:03.099142 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:27:17 crc kubenswrapper[4823]: I0320 09:27:17.097717 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:27:17 crc kubenswrapper[4823]: E0320 09:27:17.098453 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:27:32 crc kubenswrapper[4823]: I0320 09:27:32.097438 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:27:32 crc kubenswrapper[4823]: E0320 09:27:32.098133 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:27:47 crc kubenswrapper[4823]: I0320 09:27:47.097633 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:27:47 crc kubenswrapper[4823]: E0320 09:27:47.098708 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.135428 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566648-fh89w"] Mar 20 09:28:00 crc kubenswrapper[4823]: E0320 09:28:00.136505 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1591a99f-9fc4-4892-9637-298af3ea2e35" containerName="oc" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.136519 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1591a99f-9fc4-4892-9637-298af3ea2e35" containerName="oc" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.136657 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1591a99f-9fc4-4892-9637-298af3ea2e35" containerName="oc" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.137074 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566648-fh89w" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.140043 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.140575 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.140639 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.147682 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566648-fh89w"] Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.260994 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mk7j\" (UniqueName: \"kubernetes.io/projected/6dd36c8a-3914-4566-b4f4-5b3ccc186a3e-kube-api-access-6mk7j\") pod \"auto-csr-approver-29566648-fh89w\" (UID: \"6dd36c8a-3914-4566-b4f4-5b3ccc186a3e\") " pod="openshift-infra/auto-csr-approver-29566648-fh89w" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.362734 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mk7j\" (UniqueName: \"kubernetes.io/projected/6dd36c8a-3914-4566-b4f4-5b3ccc186a3e-kube-api-access-6mk7j\") pod \"auto-csr-approver-29566648-fh89w\" (UID: \"6dd36c8a-3914-4566-b4f4-5b3ccc186a3e\") " pod="openshift-infra/auto-csr-approver-29566648-fh89w" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.381795 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mk7j\" (UniqueName: \"kubernetes.io/projected/6dd36c8a-3914-4566-b4f4-5b3ccc186a3e-kube-api-access-6mk7j\") pod \"auto-csr-approver-29566648-fh89w\" (UID: \"6dd36c8a-3914-4566-b4f4-5b3ccc186a3e\") " pod="openshift-infra/auto-csr-approver-29566648-fh89w" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.457711 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566648-fh89w" Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.880735 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566648-fh89w"] Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.893262 4823 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:28:00 crc kubenswrapper[4823]: I0320 09:28:00.976044 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566648-fh89w" event={"ID":"6dd36c8a-3914-4566-b4f4-5b3ccc186a3e","Type":"ContainerStarted","Data":"fd037cffcf7dc0315cb1e601cdce1df7c42dbf5b6b32ab8fcf9f3882404f74e7"} Mar 20 09:28:02 crc kubenswrapper[4823]: I0320 09:28:02.097190 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:28:02 crc kubenswrapper[4823]: E0320 09:28:02.097722 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:28:02 crc kubenswrapper[4823]: I0320 09:28:02.990199 4823 generic.go:334] "Generic (PLEG): container finished" podID="6dd36c8a-3914-4566-b4f4-5b3ccc186a3e" containerID="04faa8f9b1799278e1dca931ecc5013c7de7dc6a77af18f558395f93e7b1b289" exitCode=0 Mar 20 09:28:02 crc kubenswrapper[4823]: I0320 09:28:02.990321 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566648-fh89w" event={"ID":"6dd36c8a-3914-4566-b4f4-5b3ccc186a3e","Type":"ContainerDied","Data":"04faa8f9b1799278e1dca931ecc5013c7de7dc6a77af18f558395f93e7b1b289"} Mar 20 09:28:04 crc kubenswrapper[4823]: I0320 09:28:04.287359 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566648-fh89w" Mar 20 09:28:04 crc kubenswrapper[4823]: I0320 09:28:04.400785 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mk7j\" (UniqueName: \"kubernetes.io/projected/6dd36c8a-3914-4566-b4f4-5b3ccc186a3e-kube-api-access-6mk7j\") pod \"6dd36c8a-3914-4566-b4f4-5b3ccc186a3e\" (UID: \"6dd36c8a-3914-4566-b4f4-5b3ccc186a3e\") " Mar 20 09:28:04 crc kubenswrapper[4823]: I0320 09:28:04.410090 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dd36c8a-3914-4566-b4f4-5b3ccc186a3e-kube-api-access-6mk7j" (OuterVolumeSpecName: "kube-api-access-6mk7j") pod "6dd36c8a-3914-4566-b4f4-5b3ccc186a3e" (UID: "6dd36c8a-3914-4566-b4f4-5b3ccc186a3e"). InnerVolumeSpecName "kube-api-access-6mk7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:28:04 crc kubenswrapper[4823]: I0320 09:28:04.503286 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mk7j\" (UniqueName: \"kubernetes.io/projected/6dd36c8a-3914-4566-b4f4-5b3ccc186a3e-kube-api-access-6mk7j\") on node \"crc\" DevicePath \"\"" Mar 20 09:28:05 crc kubenswrapper[4823]: I0320 09:28:05.020903 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566648-fh89w" event={"ID":"6dd36c8a-3914-4566-b4f4-5b3ccc186a3e","Type":"ContainerDied","Data":"fd037cffcf7dc0315cb1e601cdce1df7c42dbf5b6b32ab8fcf9f3882404f74e7"} Mar 20 09:28:05 crc kubenswrapper[4823]: I0320 09:28:05.020940 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd037cffcf7dc0315cb1e601cdce1df7c42dbf5b6b32ab8fcf9f3882404f74e7" Mar 20 09:28:05 crc kubenswrapper[4823]: I0320 09:28:05.020971 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566648-fh89w" Mar 20 09:28:05 crc kubenswrapper[4823]: I0320 09:28:05.357095 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566642-4n6qx"] Mar 20 09:28:05 crc kubenswrapper[4823]: I0320 09:28:05.365932 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566642-4n6qx"] Mar 20 09:28:06 crc kubenswrapper[4823]: I0320 09:28:06.114687 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b26860f-3bd2-4ab6-99f3-77fe0f0313eb" path="/var/lib/kubelet/pods/7b26860f-3bd2-4ab6-99f3-77fe0f0313eb/volumes" Mar 20 09:28:13 crc kubenswrapper[4823]: I0320 09:28:13.099991 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:28:13 crc kubenswrapper[4823]: E0320 09:28:13.100847 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:28:24 crc kubenswrapper[4823]: I0320 09:28:24.105155 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:28:24 crc kubenswrapper[4823]: E0320 09:28:24.106400 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:28:38 crc kubenswrapper[4823]: I0320 09:28:38.097807 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:28:38 crc kubenswrapper[4823]: E0320 09:28:38.098698 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:28:39 crc kubenswrapper[4823]: I0320 09:28:39.130846 4823 scope.go:117] "RemoveContainer" containerID="39353325502cf8e868a07ec1ac521fa943a48d36078f0b38f9c88a1f4a0f2816" Mar 20 09:28:49 crc kubenswrapper[4823]: I0320 09:28:49.097754 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:28:49 crc kubenswrapper[4823]: I0320 09:28:49.381151 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"de71ee3565ec580e57037795e1681c8c4d7337a7cc9b31a431d369a1189079d0"} Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.152538 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566650-9bvx7"] Mar 20 09:30:00 crc kubenswrapper[4823]: E0320 09:30:00.154018 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dd36c8a-3914-4566-b4f4-5b3ccc186a3e" containerName="oc" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.154049 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dd36c8a-3914-4566-b4f4-5b3ccc186a3e" containerName="oc" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.154359 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dd36c8a-3914-4566-b4f4-5b3ccc186a3e" containerName="oc" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.155175 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566650-9bvx7" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.159548 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.159796 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.160288 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.165845 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566650-9bvx7"] Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.172189 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2"] Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.174147 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.177346 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2"] Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.179951 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.180136 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.272026 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-secret-volume\") pod \"collect-profiles-29566650-6wtz2\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.272106 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-config-volume\") pod \"collect-profiles-29566650-6wtz2\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.272136 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6766b\" (UniqueName: \"kubernetes.io/projected/849b11e9-c753-4ad3-b1e0-ceb75ce78192-kube-api-access-6766b\") pod \"auto-csr-approver-29566650-9bvx7\" (UID: \"849b11e9-c753-4ad3-b1e0-ceb75ce78192\") " pod="openshift-infra/auto-csr-approver-29566650-9bvx7" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.272152 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p94h6\" (UniqueName: \"kubernetes.io/projected/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-kube-api-access-p94h6\") pod \"collect-profiles-29566650-6wtz2\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.373355 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-secret-volume\") pod \"collect-profiles-29566650-6wtz2\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.373522 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-config-volume\") pod \"collect-profiles-29566650-6wtz2\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.373571 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6766b\" (UniqueName: \"kubernetes.io/projected/849b11e9-c753-4ad3-b1e0-ceb75ce78192-kube-api-access-6766b\") pod \"auto-csr-approver-29566650-9bvx7\" (UID: \"849b11e9-c753-4ad3-b1e0-ceb75ce78192\") " pod="openshift-infra/auto-csr-approver-29566650-9bvx7" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.373606 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p94h6\" (UniqueName: \"kubernetes.io/projected/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-kube-api-access-p94h6\") pod \"collect-profiles-29566650-6wtz2\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.375438 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-config-volume\") pod \"collect-profiles-29566650-6wtz2\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.383838 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-secret-volume\") pod \"collect-profiles-29566650-6wtz2\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.392791 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6766b\" (UniqueName: \"kubernetes.io/projected/849b11e9-c753-4ad3-b1e0-ceb75ce78192-kube-api-access-6766b\") pod \"auto-csr-approver-29566650-9bvx7\" (UID: \"849b11e9-c753-4ad3-b1e0-ceb75ce78192\") " pod="openshift-infra/auto-csr-approver-29566650-9bvx7" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.405748 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p94h6\" (UniqueName: \"kubernetes.io/projected/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-kube-api-access-p94h6\") pod \"collect-profiles-29566650-6wtz2\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.515833 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566650-9bvx7" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.530528 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:00 crc kubenswrapper[4823]: I0320 09:30:00.977451 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566650-9bvx7"] Mar 20 09:30:01 crc kubenswrapper[4823]: I0320 09:30:01.043460 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2"] Mar 20 09:30:01 crc kubenswrapper[4823]: W0320 09:30:01.049974 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod080a2f8b_9ff3_46ec_a197_41c5e8d7179e.slice/crio-3ade7ed5efe57ec077484549e467fda6da60d51391f8cf6f366127123e1c83e3 WatchSource:0}: Error finding container 3ade7ed5efe57ec077484549e467fda6da60d51391f8cf6f366127123e1c83e3: Status 404 returned error can't find the container with id 3ade7ed5efe57ec077484549e467fda6da60d51391f8cf6f366127123e1c83e3 Mar 20 09:30:01 crc kubenswrapper[4823]: I0320 09:30:01.122279 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" event={"ID":"080a2f8b-9ff3-46ec-a197-41c5e8d7179e","Type":"ContainerStarted","Data":"3ade7ed5efe57ec077484549e467fda6da60d51391f8cf6f366127123e1c83e3"} Mar 20 09:30:01 crc kubenswrapper[4823]: I0320 09:30:01.123352 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566650-9bvx7" event={"ID":"849b11e9-c753-4ad3-b1e0-ceb75ce78192","Type":"ContainerStarted","Data":"c3cad5062e58675258b3fa8201dfde1c3ff8321fc06e6f4f4ba9dfdb0682e976"} Mar 20 09:30:02 crc kubenswrapper[4823]: I0320 09:30:02.133487 4823 generic.go:334] "Generic (PLEG): container finished" podID="080a2f8b-9ff3-46ec-a197-41c5e8d7179e" containerID="ce2865d6957dbbeed484b525853188ebb4c89d765182612eef191a2d11b697da" exitCode=0 Mar 20 09:30:02 crc kubenswrapper[4823]: I0320 09:30:02.133655 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" event={"ID":"080a2f8b-9ff3-46ec-a197-41c5e8d7179e","Type":"ContainerDied","Data":"ce2865d6957dbbeed484b525853188ebb4c89d765182612eef191a2d11b697da"} Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.144287 4823 generic.go:334] "Generic (PLEG): container finished" podID="849b11e9-c753-4ad3-b1e0-ceb75ce78192" containerID="e9e71b624f0fd56320ada081e9f9fef42e1930c8c63a3cddc0a09bb9b6321303" exitCode=0 Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.145009 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566650-9bvx7" event={"ID":"849b11e9-c753-4ad3-b1e0-ceb75ce78192","Type":"ContainerDied","Data":"e9e71b624f0fd56320ada081e9f9fef42e1930c8c63a3cddc0a09bb9b6321303"} Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.461698 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.531357 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-config-volume\") pod \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.531860 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p94h6\" (UniqueName: \"kubernetes.io/projected/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-kube-api-access-p94h6\") pod \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.532064 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-secret-volume\") pod \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\" (UID: \"080a2f8b-9ff3-46ec-a197-41c5e8d7179e\") " Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.532239 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-config-volume" (OuterVolumeSpecName: "config-volume") pod "080a2f8b-9ff3-46ec-a197-41c5e8d7179e" (UID: "080a2f8b-9ff3-46ec-a197-41c5e8d7179e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.532401 4823 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.537908 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-kube-api-access-p94h6" (OuterVolumeSpecName: "kube-api-access-p94h6") pod "080a2f8b-9ff3-46ec-a197-41c5e8d7179e" (UID: "080a2f8b-9ff3-46ec-a197-41c5e8d7179e"). InnerVolumeSpecName "kube-api-access-p94h6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.537941 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "080a2f8b-9ff3-46ec-a197-41c5e8d7179e" (UID: "080a2f8b-9ff3-46ec-a197-41c5e8d7179e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.633976 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p94h6\" (UniqueName: \"kubernetes.io/projected/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-kube-api-access-p94h6\") on node \"crc\" DevicePath \"\"" Mar 20 09:30:03 crc kubenswrapper[4823]: I0320 09:30:03.634041 4823 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/080a2f8b-9ff3-46ec-a197-41c5e8d7179e-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:30:04 crc kubenswrapper[4823]: I0320 09:30:04.156664 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" Mar 20 09:30:04 crc kubenswrapper[4823]: I0320 09:30:04.156993 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-6wtz2" event={"ID":"080a2f8b-9ff3-46ec-a197-41c5e8d7179e","Type":"ContainerDied","Data":"3ade7ed5efe57ec077484549e467fda6da60d51391f8cf6f366127123e1c83e3"} Mar 20 09:30:04 crc kubenswrapper[4823]: I0320 09:30:04.157074 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ade7ed5efe57ec077484549e467fda6da60d51391f8cf6f366127123e1c83e3" Mar 20 09:30:04 crc kubenswrapper[4823]: I0320 09:30:04.523424 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566650-9bvx7" Mar 20 09:30:04 crc kubenswrapper[4823]: I0320 09:30:04.652451 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6766b\" (UniqueName: \"kubernetes.io/projected/849b11e9-c753-4ad3-b1e0-ceb75ce78192-kube-api-access-6766b\") pod \"849b11e9-c753-4ad3-b1e0-ceb75ce78192\" (UID: \"849b11e9-c753-4ad3-b1e0-ceb75ce78192\") " Mar 20 09:30:04 crc kubenswrapper[4823]: I0320 09:30:04.658803 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/849b11e9-c753-4ad3-b1e0-ceb75ce78192-kube-api-access-6766b" (OuterVolumeSpecName: "kube-api-access-6766b") pod "849b11e9-c753-4ad3-b1e0-ceb75ce78192" (UID: "849b11e9-c753-4ad3-b1e0-ceb75ce78192"). InnerVolumeSpecName "kube-api-access-6766b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:30:04 crc kubenswrapper[4823]: I0320 09:30:04.755080 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6766b\" (UniqueName: \"kubernetes.io/projected/849b11e9-c753-4ad3-b1e0-ceb75ce78192-kube-api-access-6766b\") on node \"crc\" DevicePath \"\"" Mar 20 09:30:05 crc kubenswrapper[4823]: I0320 09:30:05.166203 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566650-9bvx7" event={"ID":"849b11e9-c753-4ad3-b1e0-ceb75ce78192","Type":"ContainerDied","Data":"c3cad5062e58675258b3fa8201dfde1c3ff8321fc06e6f4f4ba9dfdb0682e976"} Mar 20 09:30:05 crc kubenswrapper[4823]: I0320 09:30:05.166246 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3cad5062e58675258b3fa8201dfde1c3ff8321fc06e6f4f4ba9dfdb0682e976" Mar 20 09:30:05 crc kubenswrapper[4823]: I0320 09:30:05.166270 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566650-9bvx7" Mar 20 09:30:05 crc kubenswrapper[4823]: I0320 09:30:05.607725 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566644-xsgdd"] Mar 20 09:30:05 crc kubenswrapper[4823]: I0320 09:30:05.618740 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566644-xsgdd"] Mar 20 09:30:06 crc kubenswrapper[4823]: I0320 09:30:06.112450 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aa6bd30-bf9b-42a4-956e-ea22d0c15335" path="/var/lib/kubelet/pods/2aa6bd30-bf9b-42a4-956e-ea22d0c15335/volumes" Mar 20 09:30:39 crc kubenswrapper[4823]: I0320 09:30:39.593280 4823 scope.go:117] "RemoveContainer" containerID="65940099fccc9b3ecfacfc06580c6355bdd24d7f90213ce10826b151aa46c696" Mar 20 09:31:09 crc kubenswrapper[4823]: I0320 09:31:09.431080 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:31:09 crc kubenswrapper[4823]: I0320 09:31:09.432150 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:31:39 crc kubenswrapper[4823]: I0320 09:31:39.430589 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:31:39 crc kubenswrapper[4823]: I0320 09:31:39.433192 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.162916 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566652-75hwq"] Mar 20 09:32:00 crc kubenswrapper[4823]: E0320 09:32:00.164601 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="849b11e9-c753-4ad3-b1e0-ceb75ce78192" containerName="oc" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.164629 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="849b11e9-c753-4ad3-b1e0-ceb75ce78192" containerName="oc" Mar 20 09:32:00 crc kubenswrapper[4823]: E0320 09:32:00.164715 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="080a2f8b-9ff3-46ec-a197-41c5e8d7179e" containerName="collect-profiles" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.164731 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="080a2f8b-9ff3-46ec-a197-41c5e8d7179e" containerName="collect-profiles" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.165087 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="849b11e9-c753-4ad3-b1e0-ceb75ce78192" containerName="oc" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.165137 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="080a2f8b-9ff3-46ec-a197-41c5e8d7179e" containerName="collect-profiles" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.166018 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566652-75hwq" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.168090 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.168152 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.169610 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.171454 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566652-75hwq"] Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.350676 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztz8f\" (UniqueName: \"kubernetes.io/projected/b55af045-3b89-4ae6-a631-1d59e5803372-kube-api-access-ztz8f\") pod \"auto-csr-approver-29566652-75hwq\" (UID: \"b55af045-3b89-4ae6-a631-1d59e5803372\") " pod="openshift-infra/auto-csr-approver-29566652-75hwq" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.452846 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztz8f\" (UniqueName: \"kubernetes.io/projected/b55af045-3b89-4ae6-a631-1d59e5803372-kube-api-access-ztz8f\") pod \"auto-csr-approver-29566652-75hwq\" (UID: \"b55af045-3b89-4ae6-a631-1d59e5803372\") " pod="openshift-infra/auto-csr-approver-29566652-75hwq" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.495889 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztz8f\" (UniqueName: \"kubernetes.io/projected/b55af045-3b89-4ae6-a631-1d59e5803372-kube-api-access-ztz8f\") pod \"auto-csr-approver-29566652-75hwq\" (UID: \"b55af045-3b89-4ae6-a631-1d59e5803372\") " pod="openshift-infra/auto-csr-approver-29566652-75hwq" Mar 20 09:32:00 crc kubenswrapper[4823]: I0320 09:32:00.784771 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566652-75hwq" Mar 20 09:32:01 crc kubenswrapper[4823]: I0320 09:32:01.230342 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566652-75hwq"] Mar 20 09:32:01 crc kubenswrapper[4823]: I0320 09:32:01.584734 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566652-75hwq" event={"ID":"b55af045-3b89-4ae6-a631-1d59e5803372","Type":"ContainerStarted","Data":"4986047070ab22e92cebaef97c543fa6c24e95b44eb68427d9a6c5e188681419"} Mar 20 09:32:02 crc kubenswrapper[4823]: I0320 09:32:02.594547 4823 generic.go:334] "Generic (PLEG): container finished" podID="b55af045-3b89-4ae6-a631-1d59e5803372" containerID="7348e201d56d595384fe834f513d09d051e179bcf3626bafdfdb0e4c67f90f83" exitCode=0 Mar 20 09:32:02 crc kubenswrapper[4823]: I0320 09:32:02.594664 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566652-75hwq" event={"ID":"b55af045-3b89-4ae6-a631-1d59e5803372","Type":"ContainerDied","Data":"7348e201d56d595384fe834f513d09d051e179bcf3626bafdfdb0e4c67f90f83"} Mar 20 09:32:03 crc kubenswrapper[4823]: I0320 09:32:03.923340 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566652-75hwq" Mar 20 09:32:04 crc kubenswrapper[4823]: I0320 09:32:04.105062 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztz8f\" (UniqueName: \"kubernetes.io/projected/b55af045-3b89-4ae6-a631-1d59e5803372-kube-api-access-ztz8f\") pod \"b55af045-3b89-4ae6-a631-1d59e5803372\" (UID: \"b55af045-3b89-4ae6-a631-1d59e5803372\") " Mar 20 09:32:04 crc kubenswrapper[4823]: I0320 09:32:04.112330 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b55af045-3b89-4ae6-a631-1d59e5803372-kube-api-access-ztz8f" (OuterVolumeSpecName: "kube-api-access-ztz8f") pod "b55af045-3b89-4ae6-a631-1d59e5803372" (UID: "b55af045-3b89-4ae6-a631-1d59e5803372"). InnerVolumeSpecName "kube-api-access-ztz8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:32:04 crc kubenswrapper[4823]: I0320 09:32:04.206843 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztz8f\" (UniqueName: \"kubernetes.io/projected/b55af045-3b89-4ae6-a631-1d59e5803372-kube-api-access-ztz8f\") on node \"crc\" DevicePath \"\"" Mar 20 09:32:04 crc kubenswrapper[4823]: I0320 09:32:04.612630 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566652-75hwq" event={"ID":"b55af045-3b89-4ae6-a631-1d59e5803372","Type":"ContainerDied","Data":"4986047070ab22e92cebaef97c543fa6c24e95b44eb68427d9a6c5e188681419"} Mar 20 09:32:04 crc kubenswrapper[4823]: I0320 09:32:04.612959 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4986047070ab22e92cebaef97c543fa6c24e95b44eb68427d9a6c5e188681419" Mar 20 09:32:04 crc kubenswrapper[4823]: I0320 09:32:04.612775 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566652-75hwq" Mar 20 09:32:04 crc kubenswrapper[4823]: I0320 09:32:04.991053 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566646-gx29f"] Mar 20 09:32:04 crc kubenswrapper[4823]: I0320 09:32:04.996421 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566646-gx29f"] Mar 20 09:32:06 crc kubenswrapper[4823]: I0320 09:32:06.114225 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1591a99f-9fc4-4892-9637-298af3ea2e35" path="/var/lib/kubelet/pods/1591a99f-9fc4-4892-9637-298af3ea2e35/volumes" Mar 20 09:32:09 crc kubenswrapper[4823]: I0320 09:32:09.430996 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:32:09 crc kubenswrapper[4823]: I0320 09:32:09.431425 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:32:09 crc kubenswrapper[4823]: I0320 09:32:09.431473 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:32:09 crc kubenswrapper[4823]: I0320 09:32:09.432204 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"de71ee3565ec580e57037795e1681c8c4d7337a7cc9b31a431d369a1189079d0"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:32:09 crc kubenswrapper[4823]: I0320 09:32:09.432266 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://de71ee3565ec580e57037795e1681c8c4d7337a7cc9b31a431d369a1189079d0" gracePeriod=600 Mar 20 09:32:09 crc kubenswrapper[4823]: I0320 09:32:09.674095 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="de71ee3565ec580e57037795e1681c8c4d7337a7cc9b31a431d369a1189079d0" exitCode=0 Mar 20 09:32:09 crc kubenswrapper[4823]: I0320 09:32:09.674163 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"de71ee3565ec580e57037795e1681c8c4d7337a7cc9b31a431d369a1189079d0"} Mar 20 09:32:09 crc kubenswrapper[4823]: I0320 09:32:09.674971 4823 scope.go:117] "RemoveContainer" containerID="6571cdb136be8cb7858589e5f72e24b6a89289cc2b00b01a907374f3736596e8" Mar 20 09:32:10 crc kubenswrapper[4823]: I0320 09:32:10.682987 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0"} Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.347513 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x5pw6"] Mar 20 09:32:16 crc kubenswrapper[4823]: E0320 09:32:16.348672 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b55af045-3b89-4ae6-a631-1d59e5803372" containerName="oc" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.348694 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="b55af045-3b89-4ae6-a631-1d59e5803372" containerName="oc" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.348975 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="b55af045-3b89-4ae6-a631-1d59e5803372" containerName="oc" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.350683 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.373599 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x5pw6"] Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.484835 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-catalog-content\") pod \"redhat-operators-x5pw6\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.484909 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67dtp\" (UniqueName: \"kubernetes.io/projected/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-kube-api-access-67dtp\") pod \"redhat-operators-x5pw6\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.484990 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-utilities\") pod \"redhat-operators-x5pw6\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.585939 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-utilities\") pod \"redhat-operators-x5pw6\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.586053 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-catalog-content\") pod \"redhat-operators-x5pw6\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.586073 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67dtp\" (UniqueName: \"kubernetes.io/projected/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-kube-api-access-67dtp\") pod \"redhat-operators-x5pw6\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.586506 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-utilities\") pod \"redhat-operators-x5pw6\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.586680 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-catalog-content\") pod \"redhat-operators-x5pw6\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.610840 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67dtp\" (UniqueName: \"kubernetes.io/projected/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-kube-api-access-67dtp\") pod \"redhat-operators-x5pw6\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:16 crc kubenswrapper[4823]: I0320 09:32:16.682524 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:17 crc kubenswrapper[4823]: I0320 09:32:17.173717 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x5pw6"] Mar 20 09:32:17 crc kubenswrapper[4823]: I0320 09:32:17.739349 4823 generic.go:334] "Generic (PLEG): container finished" podID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerID="89a943ce887cf451fdfd2abd6383a5404aa9a4ebfa050ffc1852b842a0fc0f0d" exitCode=0 Mar 20 09:32:17 crc kubenswrapper[4823]: I0320 09:32:17.739668 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pw6" event={"ID":"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa","Type":"ContainerDied","Data":"89a943ce887cf451fdfd2abd6383a5404aa9a4ebfa050ffc1852b842a0fc0f0d"} Mar 20 09:32:17 crc kubenswrapper[4823]: I0320 09:32:17.739724 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pw6" event={"ID":"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa","Type":"ContainerStarted","Data":"66d4f4882e8f973a957ac02d50a3332b5e446c52f9f9f105fdbeaaafa9996a18"} Mar 20 09:32:18 crc kubenswrapper[4823]: I0320 09:32:18.748415 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pw6" event={"ID":"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa","Type":"ContainerStarted","Data":"051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267"} Mar 20 09:32:19 crc kubenswrapper[4823]: I0320 09:32:19.758402 4823 generic.go:334] "Generic (PLEG): container finished" podID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerID="051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267" exitCode=0 Mar 20 09:32:19 crc kubenswrapper[4823]: I0320 09:32:19.758520 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pw6" event={"ID":"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa","Type":"ContainerDied","Data":"051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267"} Mar 20 09:32:21 crc kubenswrapper[4823]: I0320 09:32:21.789130 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pw6" event={"ID":"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa","Type":"ContainerStarted","Data":"26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab"} Mar 20 09:32:21 crc kubenswrapper[4823]: I0320 09:32:21.829146 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x5pw6" podStartSLOduration=2.57941113 podStartE2EDuration="5.829120355s" podCreationTimestamp="2026-03-20 09:32:16 +0000 UTC" firstStartedPulling="2026-03-20 09:32:17.740720024 +0000 UTC m=+2104.426616408" lastFinishedPulling="2026-03-20 09:32:20.990429259 +0000 UTC m=+2107.676325633" observedRunningTime="2026-03-20 09:32:21.818674915 +0000 UTC m=+2108.504571299" watchObservedRunningTime="2026-03-20 09:32:21.829120355 +0000 UTC m=+2108.515016759" Mar 20 09:32:26 crc kubenswrapper[4823]: I0320 09:32:26.682947 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:26 crc kubenswrapper[4823]: I0320 09:32:26.683360 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:27 crc kubenswrapper[4823]: I0320 09:32:27.735265 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x5pw6" podUID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerName="registry-server" probeResult="failure" output=< Mar 20 09:32:27 crc kubenswrapper[4823]: timeout: failed to connect service ":50051" within 1s Mar 20 09:32:27 crc kubenswrapper[4823]: > Mar 20 09:32:36 crc kubenswrapper[4823]: I0320 09:32:36.761718 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:36 crc kubenswrapper[4823]: I0320 09:32:36.842656 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:37 crc kubenswrapper[4823]: I0320 09:32:37.014849 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x5pw6"] Mar 20 09:32:37 crc kubenswrapper[4823]: I0320 09:32:37.910083 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x5pw6" podUID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerName="registry-server" containerID="cri-o://26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab" gracePeriod=2 Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.291844 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.409011 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67dtp\" (UniqueName: \"kubernetes.io/projected/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-kube-api-access-67dtp\") pod \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.409131 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-catalog-content\") pod \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.409192 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-utilities\") pod \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\" (UID: \"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa\") " Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.410020 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-utilities" (OuterVolumeSpecName: "utilities") pod "281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" (UID: "281e47b9-7b01-4d8a-9faa-c5d722f7c2fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.410159 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.414502 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-kube-api-access-67dtp" (OuterVolumeSpecName: "kube-api-access-67dtp") pod "281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" (UID: "281e47b9-7b01-4d8a-9faa-c5d722f7c2fa"). InnerVolumeSpecName "kube-api-access-67dtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.510976 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67dtp\" (UniqueName: \"kubernetes.io/projected/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-kube-api-access-67dtp\") on node \"crc\" DevicePath \"\"" Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.529983 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" (UID: "281e47b9-7b01-4d8a-9faa-c5d722f7c2fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.612258 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.920702 4823 generic.go:334] "Generic (PLEG): container finished" podID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerID="26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab" exitCode=0 Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.920791 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pw6" event={"ID":"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa","Type":"ContainerDied","Data":"26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab"} Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.920850 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5pw6" event={"ID":"281e47b9-7b01-4d8a-9faa-c5d722f7c2fa","Type":"ContainerDied","Data":"66d4f4882e8f973a957ac02d50a3332b5e446c52f9f9f105fdbeaaafa9996a18"} Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.920884 4823 scope.go:117] "RemoveContainer" containerID="26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab" Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.920881 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5pw6" Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.950123 4823 scope.go:117] "RemoveContainer" containerID="051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267" Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.976359 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x5pw6"] Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.984561 4823 scope.go:117] "RemoveContainer" containerID="89a943ce887cf451fdfd2abd6383a5404aa9a4ebfa050ffc1852b842a0fc0f0d" Mar 20 09:32:38 crc kubenswrapper[4823]: I0320 09:32:38.987634 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x5pw6"] Mar 20 09:32:39 crc kubenswrapper[4823]: I0320 09:32:39.025310 4823 scope.go:117] "RemoveContainer" containerID="26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab" Mar 20 09:32:39 crc kubenswrapper[4823]: E0320 09:32:39.026049 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab\": container with ID starting with 26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab not found: ID does not exist" containerID="26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab" Mar 20 09:32:39 crc kubenswrapper[4823]: I0320 09:32:39.026118 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab"} err="failed to get container status \"26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab\": rpc error: code = NotFound desc = could not find container \"26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab\": container with ID starting with 26c3102f518ca94cd1affd12a4ec0ce86dd9add1e0149ba8267f4c7e572f6aab not found: ID does not exist" Mar 20 09:32:39 crc kubenswrapper[4823]: I0320 09:32:39.026165 4823 scope.go:117] "RemoveContainer" containerID="051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267" Mar 20 09:32:39 crc kubenswrapper[4823]: E0320 09:32:39.026792 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267\": container with ID starting with 051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267 not found: ID does not exist" containerID="051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267" Mar 20 09:32:39 crc kubenswrapper[4823]: I0320 09:32:39.026815 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267"} err="failed to get container status \"051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267\": rpc error: code = NotFound desc = could not find container \"051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267\": container with ID starting with 051ebcd28b15ed1ce746d070118109129e8310d39d0961d90dfc87e0cb54e267 not found: ID does not exist" Mar 20 09:32:39 crc kubenswrapper[4823]: I0320 09:32:39.026829 4823 scope.go:117] "RemoveContainer" containerID="89a943ce887cf451fdfd2abd6383a5404aa9a4ebfa050ffc1852b842a0fc0f0d" Mar 20 09:32:39 crc kubenswrapper[4823]: E0320 09:32:39.027167 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89a943ce887cf451fdfd2abd6383a5404aa9a4ebfa050ffc1852b842a0fc0f0d\": container with ID starting with 89a943ce887cf451fdfd2abd6383a5404aa9a4ebfa050ffc1852b842a0fc0f0d not found: ID does not exist" containerID="89a943ce887cf451fdfd2abd6383a5404aa9a4ebfa050ffc1852b842a0fc0f0d" Mar 20 09:32:39 crc kubenswrapper[4823]: I0320 09:32:39.027188 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89a943ce887cf451fdfd2abd6383a5404aa9a4ebfa050ffc1852b842a0fc0f0d"} err="failed to get container status \"89a943ce887cf451fdfd2abd6383a5404aa9a4ebfa050ffc1852b842a0fc0f0d\": rpc error: code = NotFound desc = could not find container \"89a943ce887cf451fdfd2abd6383a5404aa9a4ebfa050ffc1852b842a0fc0f0d\": container with ID starting with 89a943ce887cf451fdfd2abd6383a5404aa9a4ebfa050ffc1852b842a0fc0f0d not found: ID does not exist" Mar 20 09:32:39 crc kubenswrapper[4823]: I0320 09:32:39.709827 4823 scope.go:117] "RemoveContainer" containerID="e204714638a805ddb16b5444e92ae4cacd5c5dc8448c5a3beaa809db5648afee" Mar 20 09:32:40 crc kubenswrapper[4823]: I0320 09:32:40.111067 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" path="/var/lib/kubelet/pods/281e47b9-7b01-4d8a-9faa-c5d722f7c2fa/volumes" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.570789 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hxfbd"] Mar 20 09:33:10 crc kubenswrapper[4823]: E0320 09:33:10.572200 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerName="registry-server" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.572221 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerName="registry-server" Mar 20 09:33:10 crc kubenswrapper[4823]: E0320 09:33:10.572254 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerName="extract-utilities" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.572261 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerName="extract-utilities" Mar 20 09:33:10 crc kubenswrapper[4823]: E0320 09:33:10.572270 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerName="extract-content" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.572278 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerName="extract-content" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.572469 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="281e47b9-7b01-4d8a-9faa-c5d722f7c2fa" containerName="registry-server" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.574041 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.586550 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hxfbd"] Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.602147 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-catalog-content\") pod \"community-operators-hxfbd\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.602212 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-utilities\") pod \"community-operators-hxfbd\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.602400 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s9ht\" (UniqueName: \"kubernetes.io/projected/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-kube-api-access-5s9ht\") pod \"community-operators-hxfbd\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.703191 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s9ht\" (UniqueName: \"kubernetes.io/projected/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-kube-api-access-5s9ht\") pod \"community-operators-hxfbd\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.703301 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-catalog-content\") pod \"community-operators-hxfbd\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.703324 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-utilities\") pod \"community-operators-hxfbd\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.704541 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-catalog-content\") pod \"community-operators-hxfbd\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.704650 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-utilities\") pod \"community-operators-hxfbd\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.729152 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s9ht\" (UniqueName: \"kubernetes.io/projected/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-kube-api-access-5s9ht\") pod \"community-operators-hxfbd\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:10 crc kubenswrapper[4823]: I0320 09:33:10.893653 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:11 crc kubenswrapper[4823]: I0320 09:33:11.431609 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hxfbd"] Mar 20 09:33:11 crc kubenswrapper[4823]: W0320 09:33:11.435708 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5a0d5ed_31ae_4bf8_aa18_5298cc0c71c0.slice/crio-6b2e97a41f7bde79cacde32b220070cd047f7b0a61b1debd7698a0dca254201e WatchSource:0}: Error finding container 6b2e97a41f7bde79cacde32b220070cd047f7b0a61b1debd7698a0dca254201e: Status 404 returned error can't find the container with id 6b2e97a41f7bde79cacde32b220070cd047f7b0a61b1debd7698a0dca254201e Mar 20 09:33:12 crc kubenswrapper[4823]: I0320 09:33:12.259381 4823 generic.go:334] "Generic (PLEG): container finished" podID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" containerID="136ce6327593dcbab8bb6be5e80f2299e6c0f68d6768df8fb9b7dfde9efd4f58" exitCode=0 Mar 20 09:33:12 crc kubenswrapper[4823]: I0320 09:33:12.259899 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxfbd" event={"ID":"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0","Type":"ContainerDied","Data":"136ce6327593dcbab8bb6be5e80f2299e6c0f68d6768df8fb9b7dfde9efd4f58"} Mar 20 09:33:12 crc kubenswrapper[4823]: I0320 09:33:12.259942 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxfbd" event={"ID":"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0","Type":"ContainerStarted","Data":"6b2e97a41f7bde79cacde32b220070cd047f7b0a61b1debd7698a0dca254201e"} Mar 20 09:33:12 crc kubenswrapper[4823]: I0320 09:33:12.266543 4823 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:33:13 crc kubenswrapper[4823]: I0320 09:33:13.270789 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxfbd" event={"ID":"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0","Type":"ContainerStarted","Data":"66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4"} Mar 20 09:33:14 crc kubenswrapper[4823]: I0320 09:33:14.282410 4823 generic.go:334] "Generic (PLEG): container finished" podID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" containerID="66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4" exitCode=0 Mar 20 09:33:14 crc kubenswrapper[4823]: I0320 09:33:14.282631 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxfbd" event={"ID":"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0","Type":"ContainerDied","Data":"66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4"} Mar 20 09:33:15 crc kubenswrapper[4823]: I0320 09:33:15.290101 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxfbd" event={"ID":"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0","Type":"ContainerStarted","Data":"9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69"} Mar 20 09:33:15 crc kubenswrapper[4823]: I0320 09:33:15.308401 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hxfbd" podStartSLOduration=2.893230472 podStartE2EDuration="5.308379762s" podCreationTimestamp="2026-03-20 09:33:10 +0000 UTC" firstStartedPulling="2026-03-20 09:33:12.265187879 +0000 UTC m=+2158.951084303" lastFinishedPulling="2026-03-20 09:33:14.680337179 +0000 UTC m=+2161.366233593" observedRunningTime="2026-03-20 09:33:15.303348624 +0000 UTC m=+2161.989245008" watchObservedRunningTime="2026-03-20 09:33:15.308379762 +0000 UTC m=+2161.994276146" Mar 20 09:33:19 crc kubenswrapper[4823]: I0320 09:33:19.946388 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wckpm"] Mar 20 09:33:19 crc kubenswrapper[4823]: I0320 09:33:19.950939 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:19 crc kubenswrapper[4823]: I0320 09:33:19.966040 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wckpm"] Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.136097 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzkdf\" (UniqueName: \"kubernetes.io/projected/eb9ea25e-5713-481d-bccf-e1d3f846f28b-kube-api-access-xzkdf\") pod \"certified-operators-wckpm\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.136380 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-catalog-content\") pod \"certified-operators-wckpm\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.136553 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-utilities\") pod \"certified-operators-wckpm\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.238373 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-utilities\") pod \"certified-operators-wckpm\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.238491 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzkdf\" (UniqueName: \"kubernetes.io/projected/eb9ea25e-5713-481d-bccf-e1d3f846f28b-kube-api-access-xzkdf\") pod \"certified-operators-wckpm\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.238551 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-catalog-content\") pod \"certified-operators-wckpm\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.238998 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-utilities\") pod \"certified-operators-wckpm\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.239055 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-catalog-content\") pod \"certified-operators-wckpm\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.262476 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzkdf\" (UniqueName: \"kubernetes.io/projected/eb9ea25e-5713-481d-bccf-e1d3f846f28b-kube-api-access-xzkdf\") pod \"certified-operators-wckpm\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.277367 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.782653 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wckpm"] Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.894977 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.895035 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:20 crc kubenswrapper[4823]: I0320 09:33:20.944963 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:21 crc kubenswrapper[4823]: I0320 09:33:21.340804 4823 generic.go:334] "Generic (PLEG): container finished" podID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" containerID="2e1e639f5ec0f2e036067bae90cf75272040e63b09c3419c5f19ed75d7bb143e" exitCode=0 Mar 20 09:33:21 crc kubenswrapper[4823]: I0320 09:33:21.340849 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wckpm" event={"ID":"eb9ea25e-5713-481d-bccf-e1d3f846f28b","Type":"ContainerDied","Data":"2e1e639f5ec0f2e036067bae90cf75272040e63b09c3419c5f19ed75d7bb143e"} Mar 20 09:33:21 crc kubenswrapper[4823]: I0320 09:33:21.341269 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wckpm" event={"ID":"eb9ea25e-5713-481d-bccf-e1d3f846f28b","Type":"ContainerStarted","Data":"0e5f3177e55f58bcaeb92b2135e1de967fd577b35d7b369c80735db9509dbd6d"} Mar 20 09:33:21 crc kubenswrapper[4823]: I0320 09:33:21.458423 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:22 crc kubenswrapper[4823]: I0320 09:33:22.350699 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wckpm" event={"ID":"eb9ea25e-5713-481d-bccf-e1d3f846f28b","Type":"ContainerStarted","Data":"481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad"} Mar 20 09:33:23 crc kubenswrapper[4823]: I0320 09:33:23.364099 4823 generic.go:334] "Generic (PLEG): container finished" podID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" containerID="481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad" exitCode=0 Mar 20 09:33:23 crc kubenswrapper[4823]: I0320 09:33:23.364142 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wckpm" event={"ID":"eb9ea25e-5713-481d-bccf-e1d3f846f28b","Type":"ContainerDied","Data":"481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad"} Mar 20 09:33:24 crc kubenswrapper[4823]: I0320 09:33:24.375943 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wckpm" event={"ID":"eb9ea25e-5713-481d-bccf-e1d3f846f28b","Type":"ContainerStarted","Data":"ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15"} Mar 20 09:33:24 crc kubenswrapper[4823]: I0320 09:33:24.398686 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wckpm" podStartSLOduration=2.861986999 podStartE2EDuration="5.398649713s" podCreationTimestamp="2026-03-20 09:33:19 +0000 UTC" firstStartedPulling="2026-03-20 09:33:21.342776201 +0000 UTC m=+2168.028672605" lastFinishedPulling="2026-03-20 09:33:23.879438945 +0000 UTC m=+2170.565335319" observedRunningTime="2026-03-20 09:33:24.393416528 +0000 UTC m=+2171.079312942" watchObservedRunningTime="2026-03-20 09:33:24.398649713 +0000 UTC m=+2171.084546087" Mar 20 09:33:25 crc kubenswrapper[4823]: I0320 09:33:25.932880 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hxfbd"] Mar 20 09:33:25 crc kubenswrapper[4823]: I0320 09:33:25.933176 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hxfbd" podUID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" containerName="registry-server" containerID="cri-o://9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69" gracePeriod=2 Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.321658 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.402437 4823 generic.go:334] "Generic (PLEG): container finished" podID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" containerID="9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69" exitCode=0 Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.402482 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxfbd" event={"ID":"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0","Type":"ContainerDied","Data":"9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69"} Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.402508 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxfbd" event={"ID":"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0","Type":"ContainerDied","Data":"6b2e97a41f7bde79cacde32b220070cd047f7b0a61b1debd7698a0dca254201e"} Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.402525 4823 scope.go:117] "RemoveContainer" containerID="9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.402599 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxfbd" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.429326 4823 scope.go:117] "RemoveContainer" containerID="66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.431013 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-catalog-content\") pod \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.431106 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s9ht\" (UniqueName: \"kubernetes.io/projected/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-kube-api-access-5s9ht\") pod \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.431261 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-utilities\") pod \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\" (UID: \"f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0\") " Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.436144 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-utilities" (OuterVolumeSpecName: "utilities") pod "f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" (UID: "f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.439586 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-kube-api-access-5s9ht" (OuterVolumeSpecName: "kube-api-access-5s9ht") pod "f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" (UID: "f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0"). InnerVolumeSpecName "kube-api-access-5s9ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.453965 4823 scope.go:117] "RemoveContainer" containerID="136ce6327593dcbab8bb6be5e80f2299e6c0f68d6768df8fb9b7dfde9efd4f58" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.504887 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" (UID: "f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.531866 4823 scope.go:117] "RemoveContainer" containerID="9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69" Mar 20 09:33:26 crc kubenswrapper[4823]: E0320 09:33:26.532337 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69\": container with ID starting with 9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69 not found: ID does not exist" containerID="9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.532377 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69"} err="failed to get container status \"9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69\": rpc error: code = NotFound desc = could not find container \"9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69\": container with ID starting with 9a49dc6034c3570e7ca16eba6339f275f010645ba032d9de6c3c2c4c7e51ef69 not found: ID does not exist" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.532403 4823 scope.go:117] "RemoveContainer" containerID="66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4" Mar 20 09:33:26 crc kubenswrapper[4823]: E0320 09:33:26.532725 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4\": container with ID starting with 66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4 not found: ID does not exist" containerID="66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.532772 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4"} err="failed to get container status \"66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4\": rpc error: code = NotFound desc = could not find container \"66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4\": container with ID starting with 66c76a3ceeada8be853ffaf7f0012395c774c23791068d86e80c287f0935d8f4 not found: ID does not exist" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.532793 4823 scope.go:117] "RemoveContainer" containerID="136ce6327593dcbab8bb6be5e80f2299e6c0f68d6768df8fb9b7dfde9efd4f58" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.532967 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.533010 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s9ht\" (UniqueName: \"kubernetes.io/projected/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-kube-api-access-5s9ht\") on node \"crc\" DevicePath \"\"" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.533031 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:33:26 crc kubenswrapper[4823]: E0320 09:33:26.533210 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"136ce6327593dcbab8bb6be5e80f2299e6c0f68d6768df8fb9b7dfde9efd4f58\": container with ID starting with 136ce6327593dcbab8bb6be5e80f2299e6c0f68d6768df8fb9b7dfde9efd4f58 not found: ID does not exist" containerID="136ce6327593dcbab8bb6be5e80f2299e6c0f68d6768df8fb9b7dfde9efd4f58" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.533252 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"136ce6327593dcbab8bb6be5e80f2299e6c0f68d6768df8fb9b7dfde9efd4f58"} err="failed to get container status \"136ce6327593dcbab8bb6be5e80f2299e6c0f68d6768df8fb9b7dfde9efd4f58\": rpc error: code = NotFound desc = could not find container \"136ce6327593dcbab8bb6be5e80f2299e6c0f68d6768df8fb9b7dfde9efd4f58\": container with ID starting with 136ce6327593dcbab8bb6be5e80f2299e6c0f68d6768df8fb9b7dfde9efd4f58 not found: ID does not exist" Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.756434 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hxfbd"] Mar 20 09:33:26 crc kubenswrapper[4823]: I0320 09:33:26.761089 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hxfbd"] Mar 20 09:33:28 crc kubenswrapper[4823]: I0320 09:33:28.110376 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" path="/var/lib/kubelet/pods/f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0/volumes" Mar 20 09:33:30 crc kubenswrapper[4823]: I0320 09:33:30.278153 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:30 crc kubenswrapper[4823]: I0320 09:33:30.278536 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:30 crc kubenswrapper[4823]: I0320 09:33:30.336724 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:30 crc kubenswrapper[4823]: I0320 09:33:30.515924 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:31 crc kubenswrapper[4823]: I0320 09:33:31.535044 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wckpm"] Mar 20 09:33:32 crc kubenswrapper[4823]: I0320 09:33:32.454518 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wckpm" podUID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" containerName="registry-server" containerID="cri-o://ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15" gracePeriod=2 Mar 20 09:33:32 crc kubenswrapper[4823]: I0320 09:33:32.989786 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.135415 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzkdf\" (UniqueName: \"kubernetes.io/projected/eb9ea25e-5713-481d-bccf-e1d3f846f28b-kube-api-access-xzkdf\") pod \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.135478 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-catalog-content\") pod \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.135523 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-utilities\") pod \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\" (UID: \"eb9ea25e-5713-481d-bccf-e1d3f846f28b\") " Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.136848 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-utilities" (OuterVolumeSpecName: "utilities") pod "eb9ea25e-5713-481d-bccf-e1d3f846f28b" (UID: "eb9ea25e-5713-481d-bccf-e1d3f846f28b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.145628 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb9ea25e-5713-481d-bccf-e1d3f846f28b-kube-api-access-xzkdf" (OuterVolumeSpecName: "kube-api-access-xzkdf") pod "eb9ea25e-5713-481d-bccf-e1d3f846f28b" (UID: "eb9ea25e-5713-481d-bccf-e1d3f846f28b"). InnerVolumeSpecName "kube-api-access-xzkdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.238231 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzkdf\" (UniqueName: \"kubernetes.io/projected/eb9ea25e-5713-481d-bccf-e1d3f846f28b-kube-api-access-xzkdf\") on node \"crc\" DevicePath \"\"" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.238261 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.466208 4823 generic.go:334] "Generic (PLEG): container finished" podID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" containerID="ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15" exitCode=0 Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.466265 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wckpm" event={"ID":"eb9ea25e-5713-481d-bccf-e1d3f846f28b","Type":"ContainerDied","Data":"ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15"} Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.466303 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wckpm" event={"ID":"eb9ea25e-5713-481d-bccf-e1d3f846f28b","Type":"ContainerDied","Data":"0e5f3177e55f58bcaeb92b2135e1de967fd577b35d7b369c80735db9509dbd6d"} Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.466330 4823 scope.go:117] "RemoveContainer" containerID="ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.466375 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wckpm" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.496272 4823 scope.go:117] "RemoveContainer" containerID="481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.523264 4823 scope.go:117] "RemoveContainer" containerID="2e1e639f5ec0f2e036067bae90cf75272040e63b09c3419c5f19ed75d7bb143e" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.574069 4823 scope.go:117] "RemoveContainer" containerID="ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15" Mar 20 09:33:33 crc kubenswrapper[4823]: E0320 09:33:33.575241 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15\": container with ID starting with ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15 not found: ID does not exist" containerID="ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.575315 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15"} err="failed to get container status \"ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15\": rpc error: code = NotFound desc = could not find container \"ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15\": container with ID starting with ecf2ca087fdc1df91fa94b77569b1eeabd6874ba39d070b24935b0477b3bad15 not found: ID does not exist" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.575347 4823 scope.go:117] "RemoveContainer" containerID="481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad" Mar 20 09:33:33 crc kubenswrapper[4823]: E0320 09:33:33.575808 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad\": container with ID starting with 481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad not found: ID does not exist" containerID="481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.575842 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad"} err="failed to get container status \"481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad\": rpc error: code = NotFound desc = could not find container \"481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad\": container with ID starting with 481d2aae9d44966d5ab70f2a3f86e6cb707a8952ba3b2217ad948b2b614550ad not found: ID does not exist" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.575861 4823 scope.go:117] "RemoveContainer" containerID="2e1e639f5ec0f2e036067bae90cf75272040e63b09c3419c5f19ed75d7bb143e" Mar 20 09:33:33 crc kubenswrapper[4823]: E0320 09:33:33.576211 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e1e639f5ec0f2e036067bae90cf75272040e63b09c3419c5f19ed75d7bb143e\": container with ID starting with 2e1e639f5ec0f2e036067bae90cf75272040e63b09c3419c5f19ed75d7bb143e not found: ID does not exist" containerID="2e1e639f5ec0f2e036067bae90cf75272040e63b09c3419c5f19ed75d7bb143e" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.576254 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e1e639f5ec0f2e036067bae90cf75272040e63b09c3419c5f19ed75d7bb143e"} err="failed to get container status \"2e1e639f5ec0f2e036067bae90cf75272040e63b09c3419c5f19ed75d7bb143e\": rpc error: code = NotFound desc = could not find container \"2e1e639f5ec0f2e036067bae90cf75272040e63b09c3419c5f19ed75d7bb143e\": container with ID starting with 2e1e639f5ec0f2e036067bae90cf75272040e63b09c3419c5f19ed75d7bb143e not found: ID does not exist" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.688841 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb9ea25e-5713-481d-bccf-e1d3f846f28b" (UID: "eb9ea25e-5713-481d-bccf-e1d3f846f28b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.745916 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9ea25e-5713-481d-bccf-e1d3f846f28b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.802591 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wckpm"] Mar 20 09:33:33 crc kubenswrapper[4823]: I0320 09:33:33.814817 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wckpm"] Mar 20 09:33:34 crc kubenswrapper[4823]: I0320 09:33:34.110908 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" path="/var/lib/kubelet/pods/eb9ea25e-5713-481d-bccf-e1d3f846f28b/volumes" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.149612 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566654-7mbsb"] Mar 20 09:34:00 crc kubenswrapper[4823]: E0320 09:34:00.150385 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" containerName="extract-utilities" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.150400 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" containerName="extract-utilities" Mar 20 09:34:00 crc kubenswrapper[4823]: E0320 09:34:00.150418 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" containerName="registry-server" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.150426 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" containerName="registry-server" Mar 20 09:34:00 crc kubenswrapper[4823]: E0320 09:34:00.150442 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" containerName="extract-content" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.150450 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" containerName="extract-content" Mar 20 09:34:00 crc kubenswrapper[4823]: E0320 09:34:00.150463 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" containerName="extract-utilities" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.150471 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" containerName="extract-utilities" Mar 20 09:34:00 crc kubenswrapper[4823]: E0320 09:34:00.150493 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" containerName="registry-server" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.150501 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" containerName="registry-server" Mar 20 09:34:00 crc kubenswrapper[4823]: E0320 09:34:00.150512 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" containerName="extract-content" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.150519 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" containerName="extract-content" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.150691 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9ea25e-5713-481d-bccf-e1d3f846f28b" containerName="registry-server" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.150719 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a0d5ed-31ae-4bf8-aa18-5298cc0c71c0" containerName="registry-server" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.151244 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566654-7mbsb" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.153281 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.153376 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.154042 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.176604 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566654-7mbsb"] Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.197428 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw5jq\" (UniqueName: \"kubernetes.io/projected/1c557d70-fe79-45d7-b4ae-b7ae038fa39d-kube-api-access-zw5jq\") pod \"auto-csr-approver-29566654-7mbsb\" (UID: \"1c557d70-fe79-45d7-b4ae-b7ae038fa39d\") " pod="openshift-infra/auto-csr-approver-29566654-7mbsb" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.298580 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw5jq\" (UniqueName: \"kubernetes.io/projected/1c557d70-fe79-45d7-b4ae-b7ae038fa39d-kube-api-access-zw5jq\") pod \"auto-csr-approver-29566654-7mbsb\" (UID: \"1c557d70-fe79-45d7-b4ae-b7ae038fa39d\") " pod="openshift-infra/auto-csr-approver-29566654-7mbsb" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.332231 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw5jq\" (UniqueName: \"kubernetes.io/projected/1c557d70-fe79-45d7-b4ae-b7ae038fa39d-kube-api-access-zw5jq\") pod \"auto-csr-approver-29566654-7mbsb\" (UID: \"1c557d70-fe79-45d7-b4ae-b7ae038fa39d\") " pod="openshift-infra/auto-csr-approver-29566654-7mbsb" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.490347 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566654-7mbsb" Mar 20 09:34:00 crc kubenswrapper[4823]: I0320 09:34:00.979568 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566654-7mbsb"] Mar 20 09:34:01 crc kubenswrapper[4823]: I0320 09:34:01.712213 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566654-7mbsb" event={"ID":"1c557d70-fe79-45d7-b4ae-b7ae038fa39d","Type":"ContainerStarted","Data":"a16a6ab60d7af92fd40db274d0a38d4a65bee648ba5e307322fd9883d3153dd6"} Mar 20 09:34:02 crc kubenswrapper[4823]: I0320 09:34:02.723672 4823 generic.go:334] "Generic (PLEG): container finished" podID="1c557d70-fe79-45d7-b4ae-b7ae038fa39d" containerID="de5a8b39198b7a90826855663da67bb8689c06955edb6b29c94d9b5c2c8c79c7" exitCode=0 Mar 20 09:34:02 crc kubenswrapper[4823]: I0320 09:34:02.723811 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566654-7mbsb" event={"ID":"1c557d70-fe79-45d7-b4ae-b7ae038fa39d","Type":"ContainerDied","Data":"de5a8b39198b7a90826855663da67bb8689c06955edb6b29c94d9b5c2c8c79c7"} Mar 20 09:34:04 crc kubenswrapper[4823]: I0320 09:34:04.047115 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566654-7mbsb" Mar 20 09:34:04 crc kubenswrapper[4823]: I0320 09:34:04.152237 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw5jq\" (UniqueName: \"kubernetes.io/projected/1c557d70-fe79-45d7-b4ae-b7ae038fa39d-kube-api-access-zw5jq\") pod \"1c557d70-fe79-45d7-b4ae-b7ae038fa39d\" (UID: \"1c557d70-fe79-45d7-b4ae-b7ae038fa39d\") " Mar 20 09:34:04 crc kubenswrapper[4823]: I0320 09:34:04.159636 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c557d70-fe79-45d7-b4ae-b7ae038fa39d-kube-api-access-zw5jq" (OuterVolumeSpecName: "kube-api-access-zw5jq") pod "1c557d70-fe79-45d7-b4ae-b7ae038fa39d" (UID: "1c557d70-fe79-45d7-b4ae-b7ae038fa39d"). InnerVolumeSpecName "kube-api-access-zw5jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:34:04 crc kubenswrapper[4823]: I0320 09:34:04.254210 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw5jq\" (UniqueName: \"kubernetes.io/projected/1c557d70-fe79-45d7-b4ae-b7ae038fa39d-kube-api-access-zw5jq\") on node \"crc\" DevicePath \"\"" Mar 20 09:34:04 crc kubenswrapper[4823]: I0320 09:34:04.743231 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566654-7mbsb" event={"ID":"1c557d70-fe79-45d7-b4ae-b7ae038fa39d","Type":"ContainerDied","Data":"a16a6ab60d7af92fd40db274d0a38d4a65bee648ba5e307322fd9883d3153dd6"} Mar 20 09:34:04 crc kubenswrapper[4823]: I0320 09:34:04.743278 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a16a6ab60d7af92fd40db274d0a38d4a65bee648ba5e307322fd9883d3153dd6" Mar 20 09:34:04 crc kubenswrapper[4823]: I0320 09:34:04.743714 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566654-7mbsb" Mar 20 09:34:05 crc kubenswrapper[4823]: I0320 09:34:05.116809 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566648-fh89w"] Mar 20 09:34:05 crc kubenswrapper[4823]: I0320 09:34:05.127659 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566648-fh89w"] Mar 20 09:34:06 crc kubenswrapper[4823]: I0320 09:34:06.106521 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dd36c8a-3914-4566-b4f4-5b3ccc186a3e" path="/var/lib/kubelet/pods/6dd36c8a-3914-4566-b4f4-5b3ccc186a3e/volumes" Mar 20 09:34:09 crc kubenswrapper[4823]: I0320 09:34:09.430431 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:34:09 crc kubenswrapper[4823]: I0320 09:34:09.430958 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:34:39 crc kubenswrapper[4823]: I0320 09:34:39.430869 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:34:39 crc kubenswrapper[4823]: I0320 09:34:39.431503 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:34:39 crc kubenswrapper[4823]: I0320 09:34:39.850663 4823 scope.go:117] "RemoveContainer" containerID="04faa8f9b1799278e1dca931ecc5013c7de7dc6a77af18f558395f93e7b1b289" Mar 20 09:35:09 crc kubenswrapper[4823]: I0320 09:35:09.430011 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:35:09 crc kubenswrapper[4823]: I0320 09:35:09.430535 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:35:09 crc kubenswrapper[4823]: I0320 09:35:09.430596 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:35:09 crc kubenswrapper[4823]: I0320 09:35:09.431385 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:35:09 crc kubenswrapper[4823]: I0320 09:35:09.431437 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" gracePeriod=600 Mar 20 09:35:09 crc kubenswrapper[4823]: E0320 09:35:09.564146 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:35:10 crc kubenswrapper[4823]: I0320 09:35:10.277392 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" exitCode=0 Mar 20 09:35:10 crc kubenswrapper[4823]: I0320 09:35:10.277460 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0"} Mar 20 09:35:10 crc kubenswrapper[4823]: I0320 09:35:10.277506 4823 scope.go:117] "RemoveContainer" containerID="de71ee3565ec580e57037795e1681c8c4d7337a7cc9b31a431d369a1189079d0" Mar 20 09:35:10 crc kubenswrapper[4823]: I0320 09:35:10.281169 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:35:10 crc kubenswrapper[4823]: E0320 09:35:10.281564 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:35:23 crc kubenswrapper[4823]: I0320 09:35:23.096569 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:35:23 crc kubenswrapper[4823]: E0320 09:35:23.097209 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:35:37 crc kubenswrapper[4823]: I0320 09:35:37.098026 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:35:37 crc kubenswrapper[4823]: E0320 09:35:37.098926 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.294293 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6lfnr"] Mar 20 09:35:46 crc kubenswrapper[4823]: E0320 09:35:46.295275 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c557d70-fe79-45d7-b4ae-b7ae038fa39d" containerName="oc" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.295295 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c557d70-fe79-45d7-b4ae-b7ae038fa39d" containerName="oc" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.295521 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c557d70-fe79-45d7-b4ae-b7ae038fa39d" containerName="oc" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.299690 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.318601 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6lfnr"] Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.413619 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-catalog-content\") pod \"redhat-marketplace-6lfnr\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.413896 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-utilities\") pod \"redhat-marketplace-6lfnr\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.413923 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj9l7\" (UniqueName: \"kubernetes.io/projected/4dba1cd3-d2b3-4e88-88cd-677041ab87db-kube-api-access-pj9l7\") pod \"redhat-marketplace-6lfnr\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.515540 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-catalog-content\") pod \"redhat-marketplace-6lfnr\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.515912 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-utilities\") pod \"redhat-marketplace-6lfnr\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.516072 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj9l7\" (UniqueName: \"kubernetes.io/projected/4dba1cd3-d2b3-4e88-88cd-677041ab87db-kube-api-access-pj9l7\") pod \"redhat-marketplace-6lfnr\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.516225 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-catalog-content\") pod \"redhat-marketplace-6lfnr\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.516369 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-utilities\") pod \"redhat-marketplace-6lfnr\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.545647 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj9l7\" (UniqueName: \"kubernetes.io/projected/4dba1cd3-d2b3-4e88-88cd-677041ab87db-kube-api-access-pj9l7\") pod \"redhat-marketplace-6lfnr\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:46 crc kubenswrapper[4823]: I0320 09:35:46.623523 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:47 crc kubenswrapper[4823]: I0320 09:35:47.115503 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6lfnr"] Mar 20 09:35:47 crc kubenswrapper[4823]: I0320 09:35:47.632929 4823 generic.go:334] "Generic (PLEG): container finished" podID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" containerID="8d7a1386794ee6e69b02a890d4c73aa512403023e4747a26fde4360154aad301" exitCode=0 Mar 20 09:35:47 crc kubenswrapper[4823]: I0320 09:35:47.632998 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6lfnr" event={"ID":"4dba1cd3-d2b3-4e88-88cd-677041ab87db","Type":"ContainerDied","Data":"8d7a1386794ee6e69b02a890d4c73aa512403023e4747a26fde4360154aad301"} Mar 20 09:35:47 crc kubenswrapper[4823]: I0320 09:35:47.633035 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6lfnr" event={"ID":"4dba1cd3-d2b3-4e88-88cd-677041ab87db","Type":"ContainerStarted","Data":"386429131afea32f6762c734427bacc38bbd21d7c83bb52a791d001c38611964"} Mar 20 09:35:48 crc kubenswrapper[4823]: I0320 09:35:48.643995 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6lfnr" event={"ID":"4dba1cd3-d2b3-4e88-88cd-677041ab87db","Type":"ContainerStarted","Data":"46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8"} Mar 20 09:35:49 crc kubenswrapper[4823]: I0320 09:35:49.657827 4823 generic.go:334] "Generic (PLEG): container finished" podID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" containerID="46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8" exitCode=0 Mar 20 09:35:49 crc kubenswrapper[4823]: I0320 09:35:49.658170 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6lfnr" event={"ID":"4dba1cd3-d2b3-4e88-88cd-677041ab87db","Type":"ContainerDied","Data":"46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8"} Mar 20 09:35:50 crc kubenswrapper[4823]: I0320 09:35:50.678576 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6lfnr" event={"ID":"4dba1cd3-d2b3-4e88-88cd-677041ab87db","Type":"ContainerStarted","Data":"662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5"} Mar 20 09:35:50 crc kubenswrapper[4823]: I0320 09:35:50.709321 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6lfnr" podStartSLOduration=2.237272608 podStartE2EDuration="4.709290264s" podCreationTimestamp="2026-03-20 09:35:46 +0000 UTC" firstStartedPulling="2026-03-20 09:35:47.636938687 +0000 UTC m=+2314.322835091" lastFinishedPulling="2026-03-20 09:35:50.108956323 +0000 UTC m=+2316.794852747" observedRunningTime="2026-03-20 09:35:50.706408874 +0000 UTC m=+2317.392305268" watchObservedRunningTime="2026-03-20 09:35:50.709290264 +0000 UTC m=+2317.395186648" Mar 20 09:35:51 crc kubenswrapper[4823]: I0320 09:35:51.097026 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:35:51 crc kubenswrapper[4823]: E0320 09:35:51.097238 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:35:56 crc kubenswrapper[4823]: I0320 09:35:56.623682 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:56 crc kubenswrapper[4823]: I0320 09:35:56.624314 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:56 crc kubenswrapper[4823]: I0320 09:35:56.700027 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:56 crc kubenswrapper[4823]: I0320 09:35:56.791860 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:56 crc kubenswrapper[4823]: I0320 09:35:56.943319 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6lfnr"] Mar 20 09:35:58 crc kubenswrapper[4823]: I0320 09:35:58.747452 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6lfnr" podUID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" containerName="registry-server" containerID="cri-o://662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5" gracePeriod=2 Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.702686 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.756818 4823 generic.go:334] "Generic (PLEG): container finished" podID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" containerID="662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5" exitCode=0 Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.756859 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6lfnr" event={"ID":"4dba1cd3-d2b3-4e88-88cd-677041ab87db","Type":"ContainerDied","Data":"662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5"} Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.756888 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6lfnr" event={"ID":"4dba1cd3-d2b3-4e88-88cd-677041ab87db","Type":"ContainerDied","Data":"386429131afea32f6762c734427bacc38bbd21d7c83bb52a791d001c38611964"} Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.756905 4823 scope.go:117] "RemoveContainer" containerID="662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.756938 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6lfnr" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.775498 4823 scope.go:117] "RemoveContainer" containerID="46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.791714 4823 scope.go:117] "RemoveContainer" containerID="8d7a1386794ee6e69b02a890d4c73aa512403023e4747a26fde4360154aad301" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.814529 4823 scope.go:117] "RemoveContainer" containerID="662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5" Mar 20 09:35:59 crc kubenswrapper[4823]: E0320 09:35:59.815406 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5\": container with ID starting with 662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5 not found: ID does not exist" containerID="662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.815465 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5"} err="failed to get container status \"662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5\": rpc error: code = NotFound desc = could not find container \"662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5\": container with ID starting with 662f28036b7f824817821ce5d9b22d8f258fbda1f6822194d93e9aa36f679fb5 not found: ID does not exist" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.815499 4823 scope.go:117] "RemoveContainer" containerID="46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8" Mar 20 09:35:59 crc kubenswrapper[4823]: E0320 09:35:59.815954 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8\": container with ID starting with 46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8 not found: ID does not exist" containerID="46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.816021 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8"} err="failed to get container status \"46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8\": rpc error: code = NotFound desc = could not find container \"46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8\": container with ID starting with 46ea6a1429d4edcbc59d98d4bddc5c629401f045d2abf52823d0a314038e45c8 not found: ID does not exist" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.816065 4823 scope.go:117] "RemoveContainer" containerID="8d7a1386794ee6e69b02a890d4c73aa512403023e4747a26fde4360154aad301" Mar 20 09:35:59 crc kubenswrapper[4823]: E0320 09:35:59.816536 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d7a1386794ee6e69b02a890d4c73aa512403023e4747a26fde4360154aad301\": container with ID starting with 8d7a1386794ee6e69b02a890d4c73aa512403023e4747a26fde4360154aad301 not found: ID does not exist" containerID="8d7a1386794ee6e69b02a890d4c73aa512403023e4747a26fde4360154aad301" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.816587 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d7a1386794ee6e69b02a890d4c73aa512403023e4747a26fde4360154aad301"} err="failed to get container status \"8d7a1386794ee6e69b02a890d4c73aa512403023e4747a26fde4360154aad301\": rpc error: code = NotFound desc = could not find container \"8d7a1386794ee6e69b02a890d4c73aa512403023e4747a26fde4360154aad301\": container with ID starting with 8d7a1386794ee6e69b02a890d4c73aa512403023e4747a26fde4360154aad301 not found: ID does not exist" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.873267 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-catalog-content\") pod \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.873469 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-utilities\") pod \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.873557 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj9l7\" (UniqueName: \"kubernetes.io/projected/4dba1cd3-d2b3-4e88-88cd-677041ab87db-kube-api-access-pj9l7\") pod \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\" (UID: \"4dba1cd3-d2b3-4e88-88cd-677041ab87db\") " Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.875028 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-utilities" (OuterVolumeSpecName: "utilities") pod "4dba1cd3-d2b3-4e88-88cd-677041ab87db" (UID: "4dba1cd3-d2b3-4e88-88cd-677041ab87db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.880798 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dba1cd3-d2b3-4e88-88cd-677041ab87db-kube-api-access-pj9l7" (OuterVolumeSpecName: "kube-api-access-pj9l7") pod "4dba1cd3-d2b3-4e88-88cd-677041ab87db" (UID: "4dba1cd3-d2b3-4e88-88cd-677041ab87db"). InnerVolumeSpecName "kube-api-access-pj9l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.909991 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4dba1cd3-d2b3-4e88-88cd-677041ab87db" (UID: "4dba1cd3-d2b3-4e88-88cd-677041ab87db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.975470 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj9l7\" (UniqueName: \"kubernetes.io/projected/4dba1cd3-d2b3-4e88-88cd-677041ab87db-kube-api-access-pj9l7\") on node \"crc\" DevicePath \"\"" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.975527 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:35:59 crc kubenswrapper[4823]: I0320 09:35:59.975546 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dba1cd3-d2b3-4e88-88cd-677041ab87db-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.113657 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6lfnr"] Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.114919 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6lfnr"] Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.196728 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566656-xs8cb"] Mar 20 09:36:00 crc kubenswrapper[4823]: E0320 09:36:00.197227 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" containerName="registry-server" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.197240 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" containerName="registry-server" Mar 20 09:36:00 crc kubenswrapper[4823]: E0320 09:36:00.197259 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" containerName="extract-utilities" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.197266 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" containerName="extract-utilities" Mar 20 09:36:00 crc kubenswrapper[4823]: E0320 09:36:00.197287 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" containerName="extract-content" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.197293 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" containerName="extract-content" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.197424 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" containerName="registry-server" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.197850 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566656-xs8cb" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.199892 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.202719 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.202799 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.203975 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566656-xs8cb"] Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.398723 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxtcc\" (UniqueName: \"kubernetes.io/projected/78c19f2c-34ce-4c3d-b030-ad3c16d8bae9-kube-api-access-nxtcc\") pod \"auto-csr-approver-29566656-xs8cb\" (UID: \"78c19f2c-34ce-4c3d-b030-ad3c16d8bae9\") " pod="openshift-infra/auto-csr-approver-29566656-xs8cb" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.500718 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxtcc\" (UniqueName: \"kubernetes.io/projected/78c19f2c-34ce-4c3d-b030-ad3c16d8bae9-kube-api-access-nxtcc\") pod \"auto-csr-approver-29566656-xs8cb\" (UID: \"78c19f2c-34ce-4c3d-b030-ad3c16d8bae9\") " pod="openshift-infra/auto-csr-approver-29566656-xs8cb" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.526325 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxtcc\" (UniqueName: \"kubernetes.io/projected/78c19f2c-34ce-4c3d-b030-ad3c16d8bae9-kube-api-access-nxtcc\") pod \"auto-csr-approver-29566656-xs8cb\" (UID: \"78c19f2c-34ce-4c3d-b030-ad3c16d8bae9\") " pod="openshift-infra/auto-csr-approver-29566656-xs8cb" Mar 20 09:36:00 crc kubenswrapper[4823]: I0320 09:36:00.822752 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566656-xs8cb" Mar 20 09:36:01 crc kubenswrapper[4823]: I0320 09:36:01.035161 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566656-xs8cb"] Mar 20 09:36:01 crc kubenswrapper[4823]: I0320 09:36:01.776493 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566656-xs8cb" event={"ID":"78c19f2c-34ce-4c3d-b030-ad3c16d8bae9","Type":"ContainerStarted","Data":"2e9de886b4839ac2faef7ced5e84cb94ebd417cd67bed4b62dc26ddf212e1c3c"} Mar 20 09:36:02 crc kubenswrapper[4823]: I0320 09:36:02.115433 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dba1cd3-d2b3-4e88-88cd-677041ab87db" path="/var/lib/kubelet/pods/4dba1cd3-d2b3-4e88-88cd-677041ab87db/volumes" Mar 20 09:36:02 crc kubenswrapper[4823]: I0320 09:36:02.788545 4823 generic.go:334] "Generic (PLEG): container finished" podID="78c19f2c-34ce-4c3d-b030-ad3c16d8bae9" containerID="60a46619f33a931cc1ea29afb603bdd00b9fe8669b81f5abcda3fcd0753578db" exitCode=0 Mar 20 09:36:02 crc kubenswrapper[4823]: I0320 09:36:02.788640 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566656-xs8cb" event={"ID":"78c19f2c-34ce-4c3d-b030-ad3c16d8bae9","Type":"ContainerDied","Data":"60a46619f33a931cc1ea29afb603bdd00b9fe8669b81f5abcda3fcd0753578db"} Mar 20 09:36:03 crc kubenswrapper[4823]: I0320 09:36:03.097963 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:36:03 crc kubenswrapper[4823]: E0320 09:36:03.098454 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:36:04 crc kubenswrapper[4823]: I0320 09:36:04.133730 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566656-xs8cb" Mar 20 09:36:04 crc kubenswrapper[4823]: I0320 09:36:04.163603 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxtcc\" (UniqueName: \"kubernetes.io/projected/78c19f2c-34ce-4c3d-b030-ad3c16d8bae9-kube-api-access-nxtcc\") pod \"78c19f2c-34ce-4c3d-b030-ad3c16d8bae9\" (UID: \"78c19f2c-34ce-4c3d-b030-ad3c16d8bae9\") " Mar 20 09:36:04 crc kubenswrapper[4823]: I0320 09:36:04.171162 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c19f2c-34ce-4c3d-b030-ad3c16d8bae9-kube-api-access-nxtcc" (OuterVolumeSpecName: "kube-api-access-nxtcc") pod "78c19f2c-34ce-4c3d-b030-ad3c16d8bae9" (UID: "78c19f2c-34ce-4c3d-b030-ad3c16d8bae9"). InnerVolumeSpecName "kube-api-access-nxtcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:36:04 crc kubenswrapper[4823]: I0320 09:36:04.265328 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxtcc\" (UniqueName: \"kubernetes.io/projected/78c19f2c-34ce-4c3d-b030-ad3c16d8bae9-kube-api-access-nxtcc\") on node \"crc\" DevicePath \"\"" Mar 20 09:36:04 crc kubenswrapper[4823]: I0320 09:36:04.806459 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566656-xs8cb" event={"ID":"78c19f2c-34ce-4c3d-b030-ad3c16d8bae9","Type":"ContainerDied","Data":"2e9de886b4839ac2faef7ced5e84cb94ebd417cd67bed4b62dc26ddf212e1c3c"} Mar 20 09:36:04 crc kubenswrapper[4823]: I0320 09:36:04.806515 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e9de886b4839ac2faef7ced5e84cb94ebd417cd67bed4b62dc26ddf212e1c3c" Mar 20 09:36:04 crc kubenswrapper[4823]: I0320 09:36:04.806590 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566656-xs8cb" Mar 20 09:36:05 crc kubenswrapper[4823]: I0320 09:36:05.211540 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566650-9bvx7"] Mar 20 09:36:05 crc kubenswrapper[4823]: I0320 09:36:05.216302 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566650-9bvx7"] Mar 20 09:36:06 crc kubenswrapper[4823]: I0320 09:36:06.111770 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="849b11e9-c753-4ad3-b1e0-ceb75ce78192" path="/var/lib/kubelet/pods/849b11e9-c753-4ad3-b1e0-ceb75ce78192/volumes" Mar 20 09:36:17 crc kubenswrapper[4823]: I0320 09:36:17.097629 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:36:17 crc kubenswrapper[4823]: E0320 09:36:17.098473 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:36:28 crc kubenswrapper[4823]: I0320 09:36:28.100000 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:36:28 crc kubenswrapper[4823]: E0320 09:36:28.100910 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:36:39 crc kubenswrapper[4823]: I0320 09:36:39.955285 4823 scope.go:117] "RemoveContainer" containerID="e9e71b624f0fd56320ada081e9f9fef42e1930c8c63a3cddc0a09bb9b6321303" Mar 20 09:36:42 crc kubenswrapper[4823]: I0320 09:36:42.097955 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:36:42 crc kubenswrapper[4823]: E0320 09:36:42.099294 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:36:55 crc kubenswrapper[4823]: I0320 09:36:55.098270 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:36:55 crc kubenswrapper[4823]: E0320 09:36:55.099601 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:37:09 crc kubenswrapper[4823]: I0320 09:37:09.098571 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:37:09 crc kubenswrapper[4823]: E0320 09:37:09.099411 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:37:23 crc kubenswrapper[4823]: I0320 09:37:23.097538 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:37:23 crc kubenswrapper[4823]: E0320 09:37:23.098332 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:37:38 crc kubenswrapper[4823]: I0320 09:37:38.100661 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:37:38 crc kubenswrapper[4823]: E0320 09:37:38.105547 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:37:51 crc kubenswrapper[4823]: I0320 09:37:51.097257 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:37:51 crc kubenswrapper[4823]: E0320 09:37:51.098459 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.143494 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566658-kjrdn"] Mar 20 09:38:00 crc kubenswrapper[4823]: E0320 09:38:00.149491 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c19f2c-34ce-4c3d-b030-ad3c16d8bae9" containerName="oc" Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.149545 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c19f2c-34ce-4c3d-b030-ad3c16d8bae9" containerName="oc" Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.150018 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c19f2c-34ce-4c3d-b030-ad3c16d8bae9" containerName="oc" Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.151024 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566658-kjrdn" Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.151861 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566658-kjrdn"] Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.153115 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.153786 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.154127 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.293168 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljnmn\" (UniqueName: \"kubernetes.io/projected/b000bf6e-723d-4942-ba6f-7bc844eccbab-kube-api-access-ljnmn\") pod \"auto-csr-approver-29566658-kjrdn\" (UID: \"b000bf6e-723d-4942-ba6f-7bc844eccbab\") " pod="openshift-infra/auto-csr-approver-29566658-kjrdn" Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.395234 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljnmn\" (UniqueName: \"kubernetes.io/projected/b000bf6e-723d-4942-ba6f-7bc844eccbab-kube-api-access-ljnmn\") pod \"auto-csr-approver-29566658-kjrdn\" (UID: \"b000bf6e-723d-4942-ba6f-7bc844eccbab\") " pod="openshift-infra/auto-csr-approver-29566658-kjrdn" Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.427450 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljnmn\" (UniqueName: \"kubernetes.io/projected/b000bf6e-723d-4942-ba6f-7bc844eccbab-kube-api-access-ljnmn\") pod \"auto-csr-approver-29566658-kjrdn\" (UID: \"b000bf6e-723d-4942-ba6f-7bc844eccbab\") " pod="openshift-infra/auto-csr-approver-29566658-kjrdn" Mar 20 09:38:00 crc kubenswrapper[4823]: I0320 09:38:00.480608 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566658-kjrdn" Mar 20 09:38:01 crc kubenswrapper[4823]: I0320 09:38:01.489476 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566658-kjrdn"] Mar 20 09:38:01 crc kubenswrapper[4823]: I0320 09:38:01.856508 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566658-kjrdn" event={"ID":"b000bf6e-723d-4942-ba6f-7bc844eccbab","Type":"ContainerStarted","Data":"aac65234652beced3bc04fefd3f3022d1c04b1f983a2819482e125c2770ff935"} Mar 20 09:38:02 crc kubenswrapper[4823]: I0320 09:38:02.865323 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566658-kjrdn" event={"ID":"b000bf6e-723d-4942-ba6f-7bc844eccbab","Type":"ContainerStarted","Data":"d44578d0771c54df0a79918b32793016dce0a9b0ae418569e5ba01eb1fae6411"} Mar 20 09:38:02 crc kubenswrapper[4823]: I0320 09:38:02.885973 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566658-kjrdn" podStartSLOduration=1.8815201529999999 podStartE2EDuration="2.885956863s" podCreationTimestamp="2026-03-20 09:38:00 +0000 UTC" firstStartedPulling="2026-03-20 09:38:01.492578667 +0000 UTC m=+2448.178475041" lastFinishedPulling="2026-03-20 09:38:02.497015367 +0000 UTC m=+2449.182911751" observedRunningTime="2026-03-20 09:38:02.881581046 +0000 UTC m=+2449.567477420" watchObservedRunningTime="2026-03-20 09:38:02.885956863 +0000 UTC m=+2449.571853237" Mar 20 09:38:03 crc kubenswrapper[4823]: I0320 09:38:03.876577 4823 generic.go:334] "Generic (PLEG): container finished" podID="b000bf6e-723d-4942-ba6f-7bc844eccbab" containerID="d44578d0771c54df0a79918b32793016dce0a9b0ae418569e5ba01eb1fae6411" exitCode=0 Mar 20 09:38:03 crc kubenswrapper[4823]: I0320 09:38:03.876680 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566658-kjrdn" event={"ID":"b000bf6e-723d-4942-ba6f-7bc844eccbab","Type":"ContainerDied","Data":"d44578d0771c54df0a79918b32793016dce0a9b0ae418569e5ba01eb1fae6411"} Mar 20 09:38:04 crc kubenswrapper[4823]: I0320 09:38:04.107401 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:38:04 crc kubenswrapper[4823]: E0320 09:38:04.107598 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:38:05 crc kubenswrapper[4823]: I0320 09:38:05.155861 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566658-kjrdn" Mar 20 09:38:05 crc kubenswrapper[4823]: I0320 09:38:05.170472 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljnmn\" (UniqueName: \"kubernetes.io/projected/b000bf6e-723d-4942-ba6f-7bc844eccbab-kube-api-access-ljnmn\") pod \"b000bf6e-723d-4942-ba6f-7bc844eccbab\" (UID: \"b000bf6e-723d-4942-ba6f-7bc844eccbab\") " Mar 20 09:38:05 crc kubenswrapper[4823]: I0320 09:38:05.175383 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b000bf6e-723d-4942-ba6f-7bc844eccbab-kube-api-access-ljnmn" (OuterVolumeSpecName: "kube-api-access-ljnmn") pod "b000bf6e-723d-4942-ba6f-7bc844eccbab" (UID: "b000bf6e-723d-4942-ba6f-7bc844eccbab"). InnerVolumeSpecName "kube-api-access-ljnmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:38:05 crc kubenswrapper[4823]: I0320 09:38:05.272068 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljnmn\" (UniqueName: \"kubernetes.io/projected/b000bf6e-723d-4942-ba6f-7bc844eccbab-kube-api-access-ljnmn\") on node \"crc\" DevicePath \"\"" Mar 20 09:38:05 crc kubenswrapper[4823]: I0320 09:38:05.899727 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566658-kjrdn" event={"ID":"b000bf6e-723d-4942-ba6f-7bc844eccbab","Type":"ContainerDied","Data":"aac65234652beced3bc04fefd3f3022d1c04b1f983a2819482e125c2770ff935"} Mar 20 09:38:05 crc kubenswrapper[4823]: I0320 09:38:05.899824 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aac65234652beced3bc04fefd3f3022d1c04b1f983a2819482e125c2770ff935" Mar 20 09:38:05 crc kubenswrapper[4823]: I0320 09:38:05.899915 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566658-kjrdn" Mar 20 09:38:05 crc kubenswrapper[4823]: I0320 09:38:05.970737 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566652-75hwq"] Mar 20 09:38:05 crc kubenswrapper[4823]: I0320 09:38:05.978587 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566652-75hwq"] Mar 20 09:38:06 crc kubenswrapper[4823]: I0320 09:38:06.109954 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b55af045-3b89-4ae6-a631-1d59e5803372" path="/var/lib/kubelet/pods/b55af045-3b89-4ae6-a631-1d59e5803372/volumes" Mar 20 09:38:17 crc kubenswrapper[4823]: I0320 09:38:17.097948 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:38:17 crc kubenswrapper[4823]: E0320 09:38:17.098976 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:38:28 crc kubenswrapper[4823]: I0320 09:38:28.097847 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:38:28 crc kubenswrapper[4823]: E0320 09:38:28.099088 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:38:39 crc kubenswrapper[4823]: I0320 09:38:39.098050 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:38:39 crc kubenswrapper[4823]: E0320 09:38:39.099127 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:38:40 crc kubenswrapper[4823]: I0320 09:38:40.076006 4823 scope.go:117] "RemoveContainer" containerID="7348e201d56d595384fe834f513d09d051e179bcf3626bafdfdb0e4c67f90f83" Mar 20 09:38:53 crc kubenswrapper[4823]: I0320 09:38:53.097672 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:38:53 crc kubenswrapper[4823]: E0320 09:38:53.098345 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:39:06 crc kubenswrapper[4823]: I0320 09:39:06.097738 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:39:06 crc kubenswrapper[4823]: E0320 09:39:06.098847 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:39:18 crc kubenswrapper[4823]: I0320 09:39:18.097493 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:39:18 crc kubenswrapper[4823]: E0320 09:39:18.098709 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:39:32 crc kubenswrapper[4823]: I0320 09:39:32.098210 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:39:32 crc kubenswrapper[4823]: E0320 09:39:32.099465 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:39:47 crc kubenswrapper[4823]: I0320 09:39:47.098188 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:39:47 crc kubenswrapper[4823]: E0320 09:39:47.098877 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.149611 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566660-vfdkw"] Mar 20 09:40:00 crc kubenswrapper[4823]: E0320 09:40:00.150389 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b000bf6e-723d-4942-ba6f-7bc844eccbab" containerName="oc" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.150400 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="b000bf6e-723d-4942-ba6f-7bc844eccbab" containerName="oc" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.150524 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="b000bf6e-723d-4942-ba6f-7bc844eccbab" containerName="oc" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.150984 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566660-vfdkw" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.153278 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.153310 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.153442 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.154983 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566660-vfdkw"] Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.217649 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gnc6\" (UniqueName: \"kubernetes.io/projected/2cbead88-bb8c-48a1-9e89-ef0e354cc4bb-kube-api-access-7gnc6\") pod \"auto-csr-approver-29566660-vfdkw\" (UID: \"2cbead88-bb8c-48a1-9e89-ef0e354cc4bb\") " pod="openshift-infra/auto-csr-approver-29566660-vfdkw" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.318777 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gnc6\" (UniqueName: \"kubernetes.io/projected/2cbead88-bb8c-48a1-9e89-ef0e354cc4bb-kube-api-access-7gnc6\") pod \"auto-csr-approver-29566660-vfdkw\" (UID: \"2cbead88-bb8c-48a1-9e89-ef0e354cc4bb\") " pod="openshift-infra/auto-csr-approver-29566660-vfdkw" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.336158 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gnc6\" (UniqueName: \"kubernetes.io/projected/2cbead88-bb8c-48a1-9e89-ef0e354cc4bb-kube-api-access-7gnc6\") pod \"auto-csr-approver-29566660-vfdkw\" (UID: \"2cbead88-bb8c-48a1-9e89-ef0e354cc4bb\") " pod="openshift-infra/auto-csr-approver-29566660-vfdkw" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.472000 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566660-vfdkw" Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.884172 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566660-vfdkw"] Mar 20 09:40:00 crc kubenswrapper[4823]: I0320 09:40:00.893206 4823 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:40:01 crc kubenswrapper[4823]: I0320 09:40:01.097561 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:40:01 crc kubenswrapper[4823]: E0320 09:40:01.098191 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:40:01 crc kubenswrapper[4823]: I0320 09:40:01.799922 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566660-vfdkw" event={"ID":"2cbead88-bb8c-48a1-9e89-ef0e354cc4bb","Type":"ContainerStarted","Data":"b105005371d0e9460712e437c9f826206d8765d5dcca5d150ae626471914aecd"} Mar 20 09:40:02 crc kubenswrapper[4823]: I0320 09:40:02.807137 4823 generic.go:334] "Generic (PLEG): container finished" podID="2cbead88-bb8c-48a1-9e89-ef0e354cc4bb" containerID="719e58baaf61a04aa2fdb0fdd2df0380a3499ace924bf5073b6941921ed77f63" exitCode=0 Mar 20 09:40:02 crc kubenswrapper[4823]: I0320 09:40:02.807237 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566660-vfdkw" event={"ID":"2cbead88-bb8c-48a1-9e89-ef0e354cc4bb","Type":"ContainerDied","Data":"719e58baaf61a04aa2fdb0fdd2df0380a3499ace924bf5073b6941921ed77f63"} Mar 20 09:40:04 crc kubenswrapper[4823]: I0320 09:40:04.070186 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566660-vfdkw" Mar 20 09:40:04 crc kubenswrapper[4823]: I0320 09:40:04.268108 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gnc6\" (UniqueName: \"kubernetes.io/projected/2cbead88-bb8c-48a1-9e89-ef0e354cc4bb-kube-api-access-7gnc6\") pod \"2cbead88-bb8c-48a1-9e89-ef0e354cc4bb\" (UID: \"2cbead88-bb8c-48a1-9e89-ef0e354cc4bb\") " Mar 20 09:40:04 crc kubenswrapper[4823]: I0320 09:40:04.273917 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cbead88-bb8c-48a1-9e89-ef0e354cc4bb-kube-api-access-7gnc6" (OuterVolumeSpecName: "kube-api-access-7gnc6") pod "2cbead88-bb8c-48a1-9e89-ef0e354cc4bb" (UID: "2cbead88-bb8c-48a1-9e89-ef0e354cc4bb"). InnerVolumeSpecName "kube-api-access-7gnc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:40:04 crc kubenswrapper[4823]: I0320 09:40:04.371396 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gnc6\" (UniqueName: \"kubernetes.io/projected/2cbead88-bb8c-48a1-9e89-ef0e354cc4bb-kube-api-access-7gnc6\") on node \"crc\" DevicePath \"\"" Mar 20 09:40:04 crc kubenswrapper[4823]: I0320 09:40:04.826533 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566660-vfdkw" event={"ID":"2cbead88-bb8c-48a1-9e89-ef0e354cc4bb","Type":"ContainerDied","Data":"b105005371d0e9460712e437c9f826206d8765d5dcca5d150ae626471914aecd"} Mar 20 09:40:04 crc kubenswrapper[4823]: I0320 09:40:04.826597 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b105005371d0e9460712e437c9f826206d8765d5dcca5d150ae626471914aecd" Mar 20 09:40:04 crc kubenswrapper[4823]: I0320 09:40:04.826914 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566660-vfdkw" Mar 20 09:40:05 crc kubenswrapper[4823]: I0320 09:40:05.128758 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566654-7mbsb"] Mar 20 09:40:05 crc kubenswrapper[4823]: I0320 09:40:05.133974 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566654-7mbsb"] Mar 20 09:40:06 crc kubenswrapper[4823]: I0320 09:40:06.105858 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c557d70-fe79-45d7-b4ae-b7ae038fa39d" path="/var/lib/kubelet/pods/1c557d70-fe79-45d7-b4ae-b7ae038fa39d/volumes" Mar 20 09:40:12 crc kubenswrapper[4823]: I0320 09:40:12.097903 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:40:12 crc kubenswrapper[4823]: I0320 09:40:12.886709 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"5904f94da988d8d267e1ab7d3ea981d80fd558bb69192f0ba345d93f9d1d816c"} Mar 20 09:40:40 crc kubenswrapper[4823]: I0320 09:40:40.168318 4823 scope.go:117] "RemoveContainer" containerID="de5a8b39198b7a90826855663da67bb8689c06955edb6b29c94d9b5c2c8c79c7" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.148585 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566662-vzc8p"] Mar 20 09:42:00 crc kubenswrapper[4823]: E0320 09:42:00.149479 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cbead88-bb8c-48a1-9e89-ef0e354cc4bb" containerName="oc" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.149491 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cbead88-bb8c-48a1-9e89-ef0e354cc4bb" containerName="oc" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.149613 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cbead88-bb8c-48a1-9e89-ef0e354cc4bb" containerName="oc" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.150301 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566662-vzc8p" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.152892 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.153097 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.157007 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.162025 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566662-vzc8p"] Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.233116 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgwbh\" (UniqueName: \"kubernetes.io/projected/15647a49-cbe7-4a32-b42c-60c1fbb0b0ca-kube-api-access-dgwbh\") pod \"auto-csr-approver-29566662-vzc8p\" (UID: \"15647a49-cbe7-4a32-b42c-60c1fbb0b0ca\") " pod="openshift-infra/auto-csr-approver-29566662-vzc8p" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.334341 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgwbh\" (UniqueName: \"kubernetes.io/projected/15647a49-cbe7-4a32-b42c-60c1fbb0b0ca-kube-api-access-dgwbh\") pod \"auto-csr-approver-29566662-vzc8p\" (UID: \"15647a49-cbe7-4a32-b42c-60c1fbb0b0ca\") " pod="openshift-infra/auto-csr-approver-29566662-vzc8p" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.357020 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgwbh\" (UniqueName: \"kubernetes.io/projected/15647a49-cbe7-4a32-b42c-60c1fbb0b0ca-kube-api-access-dgwbh\") pod \"auto-csr-approver-29566662-vzc8p\" (UID: \"15647a49-cbe7-4a32-b42c-60c1fbb0b0ca\") " pod="openshift-infra/auto-csr-approver-29566662-vzc8p" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.473867 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566662-vzc8p" Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.714687 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566662-vzc8p"] Mar 20 09:42:00 crc kubenswrapper[4823]: W0320 09:42:00.724047 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15647a49_cbe7_4a32_b42c_60c1fbb0b0ca.slice/crio-d794b1a438b6612566d10e3ae1c4d80c87f782e6e1fe07e2e67bc63527a6dcfb WatchSource:0}: Error finding container d794b1a438b6612566d10e3ae1c4d80c87f782e6e1fe07e2e67bc63527a6dcfb: Status 404 returned error can't find the container with id d794b1a438b6612566d10e3ae1c4d80c87f782e6e1fe07e2e67bc63527a6dcfb Mar 20 09:42:00 crc kubenswrapper[4823]: I0320 09:42:00.785792 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566662-vzc8p" event={"ID":"15647a49-cbe7-4a32-b42c-60c1fbb0b0ca","Type":"ContainerStarted","Data":"d794b1a438b6612566d10e3ae1c4d80c87f782e6e1fe07e2e67bc63527a6dcfb"} Mar 20 09:42:02 crc kubenswrapper[4823]: I0320 09:42:02.803414 4823 generic.go:334] "Generic (PLEG): container finished" podID="15647a49-cbe7-4a32-b42c-60c1fbb0b0ca" containerID="ed7feeb67421feb80da3e759844f08c6a7a7abf8e357fb567a029e4fd299420f" exitCode=0 Mar 20 09:42:02 crc kubenswrapper[4823]: I0320 09:42:02.803469 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566662-vzc8p" event={"ID":"15647a49-cbe7-4a32-b42c-60c1fbb0b0ca","Type":"ContainerDied","Data":"ed7feeb67421feb80da3e759844f08c6a7a7abf8e357fb567a029e4fd299420f"} Mar 20 09:42:04 crc kubenswrapper[4823]: I0320 09:42:04.094344 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566662-vzc8p" Mar 20 09:42:04 crc kubenswrapper[4823]: I0320 09:42:04.198526 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgwbh\" (UniqueName: \"kubernetes.io/projected/15647a49-cbe7-4a32-b42c-60c1fbb0b0ca-kube-api-access-dgwbh\") pod \"15647a49-cbe7-4a32-b42c-60c1fbb0b0ca\" (UID: \"15647a49-cbe7-4a32-b42c-60c1fbb0b0ca\") " Mar 20 09:42:04 crc kubenswrapper[4823]: I0320 09:42:04.206404 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15647a49-cbe7-4a32-b42c-60c1fbb0b0ca-kube-api-access-dgwbh" (OuterVolumeSpecName: "kube-api-access-dgwbh") pod "15647a49-cbe7-4a32-b42c-60c1fbb0b0ca" (UID: "15647a49-cbe7-4a32-b42c-60c1fbb0b0ca"). InnerVolumeSpecName "kube-api-access-dgwbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:42:04 crc kubenswrapper[4823]: I0320 09:42:04.300920 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgwbh\" (UniqueName: \"kubernetes.io/projected/15647a49-cbe7-4a32-b42c-60c1fbb0b0ca-kube-api-access-dgwbh\") on node \"crc\" DevicePath \"\"" Mar 20 09:42:04 crc kubenswrapper[4823]: I0320 09:42:04.818860 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566662-vzc8p" event={"ID":"15647a49-cbe7-4a32-b42c-60c1fbb0b0ca","Type":"ContainerDied","Data":"d794b1a438b6612566d10e3ae1c4d80c87f782e6e1fe07e2e67bc63527a6dcfb"} Mar 20 09:42:04 crc kubenswrapper[4823]: I0320 09:42:04.818906 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566662-vzc8p" Mar 20 09:42:04 crc kubenswrapper[4823]: I0320 09:42:04.818921 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d794b1a438b6612566d10e3ae1c4d80c87f782e6e1fe07e2e67bc63527a6dcfb" Mar 20 09:42:05 crc kubenswrapper[4823]: I0320 09:42:05.190603 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566656-xs8cb"] Mar 20 09:42:05 crc kubenswrapper[4823]: I0320 09:42:05.201260 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566656-xs8cb"] Mar 20 09:42:06 crc kubenswrapper[4823]: I0320 09:42:06.107857 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78c19f2c-34ce-4c3d-b030-ad3c16d8bae9" path="/var/lib/kubelet/pods/78c19f2c-34ce-4c3d-b030-ad3c16d8bae9/volumes" Mar 20 09:42:39 crc kubenswrapper[4823]: I0320 09:42:39.430545 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:42:39 crc kubenswrapper[4823]: I0320 09:42:39.431199 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:42:40 crc kubenswrapper[4823]: I0320 09:42:40.258325 4823 scope.go:117] "RemoveContainer" containerID="60a46619f33a931cc1ea29afb603bdd00b9fe8669b81f5abcda3fcd0753578db" Mar 20 09:43:09 crc kubenswrapper[4823]: I0320 09:43:09.430772 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:43:09 crc kubenswrapper[4823]: I0320 09:43:09.431279 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:43:39 crc kubenswrapper[4823]: I0320 09:43:39.430791 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:43:39 crc kubenswrapper[4823]: I0320 09:43:39.431361 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:43:39 crc kubenswrapper[4823]: I0320 09:43:39.431436 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:43:39 crc kubenswrapper[4823]: I0320 09:43:39.432142 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5904f94da988d8d267e1ab7d3ea981d80fd558bb69192f0ba345d93f9d1d816c"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:43:39 crc kubenswrapper[4823]: I0320 09:43:39.432209 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://5904f94da988d8d267e1ab7d3ea981d80fd558bb69192f0ba345d93f9d1d816c" gracePeriod=600 Mar 20 09:43:39 crc kubenswrapper[4823]: I0320 09:43:39.892844 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rj6jc"] Mar 20 09:43:39 crc kubenswrapper[4823]: E0320 09:43:39.893431 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15647a49-cbe7-4a32-b42c-60c1fbb0b0ca" containerName="oc" Mar 20 09:43:39 crc kubenswrapper[4823]: I0320 09:43:39.893452 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="15647a49-cbe7-4a32-b42c-60c1fbb0b0ca" containerName="oc" Mar 20 09:43:39 crc kubenswrapper[4823]: I0320 09:43:39.893630 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="15647a49-cbe7-4a32-b42c-60c1fbb0b0ca" containerName="oc" Mar 20 09:43:39 crc kubenswrapper[4823]: I0320 09:43:39.894819 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:39 crc kubenswrapper[4823]: I0320 09:43:39.915664 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rj6jc"] Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.062432 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-catalog-content\") pod \"certified-operators-rj6jc\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.062488 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2wdv\" (UniqueName: \"kubernetes.io/projected/946c1426-03cb-4201-b3b8-c43276472480-kube-api-access-z2wdv\") pod \"certified-operators-rj6jc\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.062521 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-utilities\") pod \"certified-operators-rj6jc\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.164093 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-catalog-content\") pod \"certified-operators-rj6jc\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.164164 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2wdv\" (UniqueName: \"kubernetes.io/projected/946c1426-03cb-4201-b3b8-c43276472480-kube-api-access-z2wdv\") pod \"certified-operators-rj6jc\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.164207 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-utilities\") pod \"certified-operators-rj6jc\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.165217 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-catalog-content\") pod \"certified-operators-rj6jc\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.165221 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-utilities\") pod \"certified-operators-rj6jc\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.183897 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2wdv\" (UniqueName: \"kubernetes.io/projected/946c1426-03cb-4201-b3b8-c43276472480-kube-api-access-z2wdv\") pod \"certified-operators-rj6jc\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.217147 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.543714 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="5904f94da988d8d267e1ab7d3ea981d80fd558bb69192f0ba345d93f9d1d816c" exitCode=0 Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.543780 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"5904f94da988d8d267e1ab7d3ea981d80fd558bb69192f0ba345d93f9d1d816c"} Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.543846 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d"} Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.543866 4823 scope.go:117] "RemoveContainer" containerID="e8205d0253b57bdbe255668e0d3e4e7e1fdf5a279653da51d2b5b982d81494a0" Mar 20 09:43:40 crc kubenswrapper[4823]: I0320 09:43:40.731845 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rj6jc"] Mar 20 09:43:41 crc kubenswrapper[4823]: I0320 09:43:41.557082 4823 generic.go:334] "Generic (PLEG): container finished" podID="946c1426-03cb-4201-b3b8-c43276472480" containerID="643b5b834f1e310b17a7fab7c0c82c959b915fe49a261720337faaa1a4830d98" exitCode=0 Mar 20 09:43:41 crc kubenswrapper[4823]: I0320 09:43:41.557188 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj6jc" event={"ID":"946c1426-03cb-4201-b3b8-c43276472480","Type":"ContainerDied","Data":"643b5b834f1e310b17a7fab7c0c82c959b915fe49a261720337faaa1a4830d98"} Mar 20 09:43:41 crc kubenswrapper[4823]: I0320 09:43:41.557606 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj6jc" event={"ID":"946c1426-03cb-4201-b3b8-c43276472480","Type":"ContainerStarted","Data":"88e8b9694b16e4ce8bf2ce9fea84b292c8c0635067e44ce1075cdc08ace25d80"} Mar 20 09:43:42 crc kubenswrapper[4823]: I0320 09:43:42.567327 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj6jc" event={"ID":"946c1426-03cb-4201-b3b8-c43276472480","Type":"ContainerStarted","Data":"5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0"} Mar 20 09:43:43 crc kubenswrapper[4823]: I0320 09:43:43.575011 4823 generic.go:334] "Generic (PLEG): container finished" podID="946c1426-03cb-4201-b3b8-c43276472480" containerID="5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0" exitCode=0 Mar 20 09:43:43 crc kubenswrapper[4823]: I0320 09:43:43.575059 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj6jc" event={"ID":"946c1426-03cb-4201-b3b8-c43276472480","Type":"ContainerDied","Data":"5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0"} Mar 20 09:43:44 crc kubenswrapper[4823]: I0320 09:43:44.584165 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj6jc" event={"ID":"946c1426-03cb-4201-b3b8-c43276472480","Type":"ContainerStarted","Data":"88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee"} Mar 20 09:43:44 crc kubenswrapper[4823]: I0320 09:43:44.602369 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rj6jc" podStartSLOduration=3.150521941 podStartE2EDuration="5.602353562s" podCreationTimestamp="2026-03-20 09:43:39 +0000 UTC" firstStartedPulling="2026-03-20 09:43:41.560208892 +0000 UTC m=+2788.246105286" lastFinishedPulling="2026-03-20 09:43:44.012040533 +0000 UTC m=+2790.697936907" observedRunningTime="2026-03-20 09:43:44.602301801 +0000 UTC m=+2791.288198175" watchObservedRunningTime="2026-03-20 09:43:44.602353562 +0000 UTC m=+2791.288249936" Mar 20 09:43:50 crc kubenswrapper[4823]: I0320 09:43:50.217635 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:50 crc kubenswrapper[4823]: I0320 09:43:50.217982 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:50 crc kubenswrapper[4823]: I0320 09:43:50.259504 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:50 crc kubenswrapper[4823]: I0320 09:43:50.672065 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:50 crc kubenswrapper[4823]: I0320 09:43:50.723115 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rj6jc"] Mar 20 09:43:52 crc kubenswrapper[4823]: I0320 09:43:52.655106 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rj6jc" podUID="946c1426-03cb-4201-b3b8-c43276472480" containerName="registry-server" containerID="cri-o://88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee" gracePeriod=2 Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.084870 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.249424 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2wdv\" (UniqueName: \"kubernetes.io/projected/946c1426-03cb-4201-b3b8-c43276472480-kube-api-access-z2wdv\") pod \"946c1426-03cb-4201-b3b8-c43276472480\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.249484 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-catalog-content\") pod \"946c1426-03cb-4201-b3b8-c43276472480\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.249556 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-utilities\") pod \"946c1426-03cb-4201-b3b8-c43276472480\" (UID: \"946c1426-03cb-4201-b3b8-c43276472480\") " Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.253929 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-utilities" (OuterVolumeSpecName: "utilities") pod "946c1426-03cb-4201-b3b8-c43276472480" (UID: "946c1426-03cb-4201-b3b8-c43276472480"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.263974 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946c1426-03cb-4201-b3b8-c43276472480-kube-api-access-z2wdv" (OuterVolumeSpecName: "kube-api-access-z2wdv") pod "946c1426-03cb-4201-b3b8-c43276472480" (UID: "946c1426-03cb-4201-b3b8-c43276472480"). InnerVolumeSpecName "kube-api-access-z2wdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.351265 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2wdv\" (UniqueName: \"kubernetes.io/projected/946c1426-03cb-4201-b3b8-c43276472480-kube-api-access-z2wdv\") on node \"crc\" DevicePath \"\"" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.351298 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.666566 4823 generic.go:334] "Generic (PLEG): container finished" podID="946c1426-03cb-4201-b3b8-c43276472480" containerID="88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee" exitCode=0 Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.666625 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj6jc" event={"ID":"946c1426-03cb-4201-b3b8-c43276472480","Type":"ContainerDied","Data":"88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee"} Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.666676 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj6jc" event={"ID":"946c1426-03cb-4201-b3b8-c43276472480","Type":"ContainerDied","Data":"88e8b9694b16e4ce8bf2ce9fea84b292c8c0635067e44ce1075cdc08ace25d80"} Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.666703 4823 scope.go:117] "RemoveContainer" containerID="88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.666697 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj6jc" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.686099 4823 scope.go:117] "RemoveContainer" containerID="5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.709175 4823 scope.go:117] "RemoveContainer" containerID="643b5b834f1e310b17a7fab7c0c82c959b915fe49a261720337faaa1a4830d98" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.736305 4823 scope.go:117] "RemoveContainer" containerID="88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee" Mar 20 09:43:53 crc kubenswrapper[4823]: E0320 09:43:53.737232 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee\": container with ID starting with 88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee not found: ID does not exist" containerID="88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.737300 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee"} err="failed to get container status \"88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee\": rpc error: code = NotFound desc = could not find container \"88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee\": container with ID starting with 88609f44db37990c88e8a666d88e78a444a1d4d2d81199ea1cef510e0de7f0ee not found: ID does not exist" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.737339 4823 scope.go:117] "RemoveContainer" containerID="5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0" Mar 20 09:43:53 crc kubenswrapper[4823]: E0320 09:43:53.739256 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0\": container with ID starting with 5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0 not found: ID does not exist" containerID="5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.739329 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0"} err="failed to get container status \"5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0\": rpc error: code = NotFound desc = could not find container \"5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0\": container with ID starting with 5478525769aa60553a86aaa593ab1258369ee4d263c9c5a86aee6cbc2c5fe8d0 not found: ID does not exist" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.739380 4823 scope.go:117] "RemoveContainer" containerID="643b5b834f1e310b17a7fab7c0c82c959b915fe49a261720337faaa1a4830d98" Mar 20 09:43:53 crc kubenswrapper[4823]: E0320 09:43:53.739965 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"643b5b834f1e310b17a7fab7c0c82c959b915fe49a261720337faaa1a4830d98\": container with ID starting with 643b5b834f1e310b17a7fab7c0c82c959b915fe49a261720337faaa1a4830d98 not found: ID does not exist" containerID="643b5b834f1e310b17a7fab7c0c82c959b915fe49a261720337faaa1a4830d98" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.740140 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"643b5b834f1e310b17a7fab7c0c82c959b915fe49a261720337faaa1a4830d98"} err="failed to get container status \"643b5b834f1e310b17a7fab7c0c82c959b915fe49a261720337faaa1a4830d98\": rpc error: code = NotFound desc = could not find container \"643b5b834f1e310b17a7fab7c0c82c959b915fe49a261720337faaa1a4830d98\": container with ID starting with 643b5b834f1e310b17a7fab7c0c82c959b915fe49a261720337faaa1a4830d98 not found: ID does not exist" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.844148 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "946c1426-03cb-4201-b3b8-c43276472480" (UID: "946c1426-03cb-4201-b3b8-c43276472480"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:43:53 crc kubenswrapper[4823]: I0320 09:43:53.857767 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/946c1426-03cb-4201-b3b8-c43276472480-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:43:54 crc kubenswrapper[4823]: I0320 09:43:54.000093 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rj6jc"] Mar 20 09:43:54 crc kubenswrapper[4823]: I0320 09:43:54.006035 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rj6jc"] Mar 20 09:43:54 crc kubenswrapper[4823]: I0320 09:43:54.105553 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="946c1426-03cb-4201-b3b8-c43276472480" path="/var/lib/kubelet/pods/946c1426-03cb-4201-b3b8-c43276472480/volumes" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.143370 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566664-cb7px"] Mar 20 09:44:00 crc kubenswrapper[4823]: E0320 09:44:00.144105 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946c1426-03cb-4201-b3b8-c43276472480" containerName="extract-content" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.144120 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="946c1426-03cb-4201-b3b8-c43276472480" containerName="extract-content" Mar 20 09:44:00 crc kubenswrapper[4823]: E0320 09:44:00.144149 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946c1426-03cb-4201-b3b8-c43276472480" containerName="extract-utilities" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.144157 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="946c1426-03cb-4201-b3b8-c43276472480" containerName="extract-utilities" Mar 20 09:44:00 crc kubenswrapper[4823]: E0320 09:44:00.144173 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946c1426-03cb-4201-b3b8-c43276472480" containerName="registry-server" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.144182 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="946c1426-03cb-4201-b3b8-c43276472480" containerName="registry-server" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.144344 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="946c1426-03cb-4201-b3b8-c43276472480" containerName="registry-server" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.144878 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566664-cb7px" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.147149 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.147691 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.148976 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.150873 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566664-cb7px"] Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.156319 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpkg9\" (UniqueName: \"kubernetes.io/projected/eff8c9ec-4a29-4fdb-bb9f-375e6d48151d-kube-api-access-rpkg9\") pod \"auto-csr-approver-29566664-cb7px\" (UID: \"eff8c9ec-4a29-4fdb-bb9f-375e6d48151d\") " pod="openshift-infra/auto-csr-approver-29566664-cb7px" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.257661 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpkg9\" (UniqueName: \"kubernetes.io/projected/eff8c9ec-4a29-4fdb-bb9f-375e6d48151d-kube-api-access-rpkg9\") pod \"auto-csr-approver-29566664-cb7px\" (UID: \"eff8c9ec-4a29-4fdb-bb9f-375e6d48151d\") " pod="openshift-infra/auto-csr-approver-29566664-cb7px" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.276094 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpkg9\" (UniqueName: \"kubernetes.io/projected/eff8c9ec-4a29-4fdb-bb9f-375e6d48151d-kube-api-access-rpkg9\") pod \"auto-csr-approver-29566664-cb7px\" (UID: \"eff8c9ec-4a29-4fdb-bb9f-375e6d48151d\") " pod="openshift-infra/auto-csr-approver-29566664-cb7px" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.464122 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566664-cb7px" Mar 20 09:44:00 crc kubenswrapper[4823]: I0320 09:44:00.880605 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566664-cb7px"] Mar 20 09:44:01 crc kubenswrapper[4823]: I0320 09:44:01.725203 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566664-cb7px" event={"ID":"eff8c9ec-4a29-4fdb-bb9f-375e6d48151d","Type":"ContainerStarted","Data":"0560cb7c43718edc33ebc1e72f076522d739264f6b8196c4f4a4ba6ed59de3a8"} Mar 20 09:44:02 crc kubenswrapper[4823]: I0320 09:44:02.737288 4823 generic.go:334] "Generic (PLEG): container finished" podID="eff8c9ec-4a29-4fdb-bb9f-375e6d48151d" containerID="5254cdd127841f3ab42734cd7833c644b1cdf0e81541365cac2f53ac48b9c09e" exitCode=0 Mar 20 09:44:02 crc kubenswrapper[4823]: I0320 09:44:02.737369 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566664-cb7px" event={"ID":"eff8c9ec-4a29-4fdb-bb9f-375e6d48151d","Type":"ContainerDied","Data":"5254cdd127841f3ab42734cd7833c644b1cdf0e81541365cac2f53ac48b9c09e"} Mar 20 09:44:04 crc kubenswrapper[4823]: I0320 09:44:04.029570 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566664-cb7px" Mar 20 09:44:04 crc kubenswrapper[4823]: I0320 09:44:04.109708 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpkg9\" (UniqueName: \"kubernetes.io/projected/eff8c9ec-4a29-4fdb-bb9f-375e6d48151d-kube-api-access-rpkg9\") pod \"eff8c9ec-4a29-4fdb-bb9f-375e6d48151d\" (UID: \"eff8c9ec-4a29-4fdb-bb9f-375e6d48151d\") " Mar 20 09:44:04 crc kubenswrapper[4823]: I0320 09:44:04.115406 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff8c9ec-4a29-4fdb-bb9f-375e6d48151d-kube-api-access-rpkg9" (OuterVolumeSpecName: "kube-api-access-rpkg9") pod "eff8c9ec-4a29-4fdb-bb9f-375e6d48151d" (UID: "eff8c9ec-4a29-4fdb-bb9f-375e6d48151d"). InnerVolumeSpecName "kube-api-access-rpkg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:44:04 crc kubenswrapper[4823]: I0320 09:44:04.211203 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpkg9\" (UniqueName: \"kubernetes.io/projected/eff8c9ec-4a29-4fdb-bb9f-375e6d48151d-kube-api-access-rpkg9\") on node \"crc\" DevicePath \"\"" Mar 20 09:44:04 crc kubenswrapper[4823]: I0320 09:44:04.756238 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566664-cb7px" event={"ID":"eff8c9ec-4a29-4fdb-bb9f-375e6d48151d","Type":"ContainerDied","Data":"0560cb7c43718edc33ebc1e72f076522d739264f6b8196c4f4a4ba6ed59de3a8"} Mar 20 09:44:04 crc kubenswrapper[4823]: I0320 09:44:04.756274 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0560cb7c43718edc33ebc1e72f076522d739264f6b8196c4f4a4ba6ed59de3a8" Mar 20 09:44:04 crc kubenswrapper[4823]: I0320 09:44:04.756303 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566664-cb7px" Mar 20 09:44:05 crc kubenswrapper[4823]: I0320 09:44:05.091834 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566658-kjrdn"] Mar 20 09:44:05 crc kubenswrapper[4823]: I0320 09:44:05.098000 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566658-kjrdn"] Mar 20 09:44:06 crc kubenswrapper[4823]: I0320 09:44:06.109498 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b000bf6e-723d-4942-ba6f-7bc844eccbab" path="/var/lib/kubelet/pods/b000bf6e-723d-4942-ba6f-7bc844eccbab/volumes" Mar 20 09:44:40 crc kubenswrapper[4823]: I0320 09:44:40.329261 4823 scope.go:117] "RemoveContainer" containerID="d44578d0771c54df0a79918b32793016dce0a9b0ae418569e5ba01eb1fae6411" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.153196 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj"] Mar 20 09:45:00 crc kubenswrapper[4823]: E0320 09:45:00.154353 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff8c9ec-4a29-4fdb-bb9f-375e6d48151d" containerName="oc" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.154371 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff8c9ec-4a29-4fdb-bb9f-375e6d48151d" containerName="oc" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.154521 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff8c9ec-4a29-4fdb-bb9f-375e6d48151d" containerName="oc" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.155123 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.157205 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.160649 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.161570 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj"] Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.295807 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjmm4\" (UniqueName: \"kubernetes.io/projected/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-kube-api-access-cjmm4\") pod \"collect-profiles-29566665-gwljj\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.295960 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-secret-volume\") pod \"collect-profiles-29566665-gwljj\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.295996 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-config-volume\") pod \"collect-profiles-29566665-gwljj\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.398170 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjmm4\" (UniqueName: \"kubernetes.io/projected/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-kube-api-access-cjmm4\") pod \"collect-profiles-29566665-gwljj\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.398523 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-secret-volume\") pod \"collect-profiles-29566665-gwljj\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.398555 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-config-volume\") pod \"collect-profiles-29566665-gwljj\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.399426 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-config-volume\") pod \"collect-profiles-29566665-gwljj\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.407409 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-secret-volume\") pod \"collect-profiles-29566665-gwljj\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.413638 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjmm4\" (UniqueName: \"kubernetes.io/projected/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-kube-api-access-cjmm4\") pod \"collect-profiles-29566665-gwljj\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.481733 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:00 crc kubenswrapper[4823]: I0320 09:45:00.727822 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj"] Mar 20 09:45:01 crc kubenswrapper[4823]: I0320 09:45:01.163684 4823 generic.go:334] "Generic (PLEG): container finished" podID="e260f827-ef58-4ac7-97d7-c2cfaa584d2e" containerID="038a70f1e89511ec31a4eb74d02cc17d50d42786e1e956d0c26dfc20abd6eda1" exitCode=0 Mar 20 09:45:01 crc kubenswrapper[4823]: I0320 09:45:01.163750 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" event={"ID":"e260f827-ef58-4ac7-97d7-c2cfaa584d2e","Type":"ContainerDied","Data":"038a70f1e89511ec31a4eb74d02cc17d50d42786e1e956d0c26dfc20abd6eda1"} Mar 20 09:45:01 crc kubenswrapper[4823]: I0320 09:45:01.163808 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" event={"ID":"e260f827-ef58-4ac7-97d7-c2cfaa584d2e","Type":"ContainerStarted","Data":"f56eba3a624d5c72e2a354e6225f7265c38ed8c41046e37d95afc4f123432c57"} Mar 20 09:45:02 crc kubenswrapper[4823]: I0320 09:45:02.518223 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:02 crc kubenswrapper[4823]: I0320 09:45:02.530196 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-config-volume\") pod \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " Mar 20 09:45:02 crc kubenswrapper[4823]: I0320 09:45:02.530576 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-secret-volume\") pod \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " Mar 20 09:45:02 crc kubenswrapper[4823]: I0320 09:45:02.530666 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjmm4\" (UniqueName: \"kubernetes.io/projected/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-kube-api-access-cjmm4\") pod \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\" (UID: \"e260f827-ef58-4ac7-97d7-c2cfaa584d2e\") " Mar 20 09:45:02 crc kubenswrapper[4823]: I0320 09:45:02.531991 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-config-volume" (OuterVolumeSpecName: "config-volume") pod "e260f827-ef58-4ac7-97d7-c2cfaa584d2e" (UID: "e260f827-ef58-4ac7-97d7-c2cfaa584d2e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:45:02 crc kubenswrapper[4823]: I0320 09:45:02.537787 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-kube-api-access-cjmm4" (OuterVolumeSpecName: "kube-api-access-cjmm4") pod "e260f827-ef58-4ac7-97d7-c2cfaa584d2e" (UID: "e260f827-ef58-4ac7-97d7-c2cfaa584d2e"). InnerVolumeSpecName "kube-api-access-cjmm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:45:02 crc kubenswrapper[4823]: I0320 09:45:02.537940 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e260f827-ef58-4ac7-97d7-c2cfaa584d2e" (UID: "e260f827-ef58-4ac7-97d7-c2cfaa584d2e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:45:02 crc kubenswrapper[4823]: I0320 09:45:02.634043 4823 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:02 crc kubenswrapper[4823]: I0320 09:45:02.634113 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjmm4\" (UniqueName: \"kubernetes.io/projected/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-kube-api-access-cjmm4\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:02 crc kubenswrapper[4823]: I0320 09:45:02.634126 4823 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e260f827-ef58-4ac7-97d7-c2cfaa584d2e-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:03 crc kubenswrapper[4823]: I0320 09:45:03.180891 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" event={"ID":"e260f827-ef58-4ac7-97d7-c2cfaa584d2e","Type":"ContainerDied","Data":"f56eba3a624d5c72e2a354e6225f7265c38ed8c41046e37d95afc4f123432c57"} Mar 20 09:45:03 crc kubenswrapper[4823]: I0320 09:45:03.181193 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f56eba3a624d5c72e2a354e6225f7265c38ed8c41046e37d95afc4f123432c57" Mar 20 09:45:03 crc kubenswrapper[4823]: I0320 09:45:03.180951 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-gwljj" Mar 20 09:45:03 crc kubenswrapper[4823]: I0320 09:45:03.598622 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8"] Mar 20 09:45:03 crc kubenswrapper[4823]: I0320 09:45:03.603924 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566620-smfx8"] Mar 20 09:45:04 crc kubenswrapper[4823]: I0320 09:45:04.111573 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7" path="/var/lib/kubelet/pods/4371c4f7-4bd9-4042-a0d2-f0348e9f0cd7/volumes" Mar 20 09:45:39 crc kubenswrapper[4823]: I0320 09:45:39.430002 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:45:39 crc kubenswrapper[4823]: I0320 09:45:39.431749 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:45:40 crc kubenswrapper[4823]: I0320 09:45:40.395379 4823 scope.go:117] "RemoveContainer" containerID="8262abc138eac431152da2dcf732d57c8026e0ca34981101a5a1495833ef3aff" Mar 20 09:45:47 crc kubenswrapper[4823]: I0320 09:45:47.951022 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d86qb"] Mar 20 09:45:47 crc kubenswrapper[4823]: E0320 09:45:47.951638 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e260f827-ef58-4ac7-97d7-c2cfaa584d2e" containerName="collect-profiles" Mar 20 09:45:47 crc kubenswrapper[4823]: I0320 09:45:47.951649 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="e260f827-ef58-4ac7-97d7-c2cfaa584d2e" containerName="collect-profiles" Mar 20 09:45:47 crc kubenswrapper[4823]: I0320 09:45:47.951800 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="e260f827-ef58-4ac7-97d7-c2cfaa584d2e" containerName="collect-profiles" Mar 20 09:45:47 crc kubenswrapper[4823]: I0320 09:45:47.952661 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:47 crc kubenswrapper[4823]: I0320 09:45:47.962450 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d86qb"] Mar 20 09:45:48 crc kubenswrapper[4823]: I0320 09:45:48.004146 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-catalog-content\") pod \"redhat-marketplace-d86qb\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:48 crc kubenswrapper[4823]: I0320 09:45:48.004234 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96759\" (UniqueName: \"kubernetes.io/projected/7ca833ad-eed8-458f-b348-b488623a3525-kube-api-access-96759\") pod \"redhat-marketplace-d86qb\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:48 crc kubenswrapper[4823]: I0320 09:45:48.004300 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-utilities\") pod \"redhat-marketplace-d86qb\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:48 crc kubenswrapper[4823]: I0320 09:45:48.105087 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-utilities\") pod \"redhat-marketplace-d86qb\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:48 crc kubenswrapper[4823]: I0320 09:45:48.105143 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-catalog-content\") pod \"redhat-marketplace-d86qb\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:48 crc kubenswrapper[4823]: I0320 09:45:48.105197 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96759\" (UniqueName: \"kubernetes.io/projected/7ca833ad-eed8-458f-b348-b488623a3525-kube-api-access-96759\") pod \"redhat-marketplace-d86qb\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:48 crc kubenswrapper[4823]: I0320 09:45:48.105795 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-utilities\") pod \"redhat-marketplace-d86qb\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:48 crc kubenswrapper[4823]: I0320 09:45:48.105890 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-catalog-content\") pod \"redhat-marketplace-d86qb\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:48 crc kubenswrapper[4823]: I0320 09:45:48.130791 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96759\" (UniqueName: \"kubernetes.io/projected/7ca833ad-eed8-458f-b348-b488623a3525-kube-api-access-96759\") pod \"redhat-marketplace-d86qb\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:48 crc kubenswrapper[4823]: I0320 09:45:48.275009 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:48 crc kubenswrapper[4823]: I0320 09:45:48.520633 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d86qb"] Mar 20 09:45:49 crc kubenswrapper[4823]: I0320 09:45:49.531583 4823 generic.go:334] "Generic (PLEG): container finished" podID="7ca833ad-eed8-458f-b348-b488623a3525" containerID="e1ee985454c08de0b977c6cb71ccd92810e2eb117243f3929eec4e952745314d" exitCode=0 Mar 20 09:45:49 crc kubenswrapper[4823]: I0320 09:45:49.531669 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d86qb" event={"ID":"7ca833ad-eed8-458f-b348-b488623a3525","Type":"ContainerDied","Data":"e1ee985454c08de0b977c6cb71ccd92810e2eb117243f3929eec4e952745314d"} Mar 20 09:45:49 crc kubenswrapper[4823]: I0320 09:45:49.532022 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d86qb" event={"ID":"7ca833ad-eed8-458f-b348-b488623a3525","Type":"ContainerStarted","Data":"54a9e29f295006de6ca3d0dfda1be898cb43c05fc629fbd3703a76fe2db5ca83"} Mar 20 09:45:49 crc kubenswrapper[4823]: I0320 09:45:49.533799 4823 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:45:51 crc kubenswrapper[4823]: I0320 09:45:51.546640 4823 generic.go:334] "Generic (PLEG): container finished" podID="7ca833ad-eed8-458f-b348-b488623a3525" containerID="8188ec05baa2555aa0f8a3c4fb293e1cf7b8ad94d33cbe9ca2c4a1d0738e91b7" exitCode=0 Mar 20 09:45:51 crc kubenswrapper[4823]: I0320 09:45:51.546808 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d86qb" event={"ID":"7ca833ad-eed8-458f-b348-b488623a3525","Type":"ContainerDied","Data":"8188ec05baa2555aa0f8a3c4fb293e1cf7b8ad94d33cbe9ca2c4a1d0738e91b7"} Mar 20 09:45:52 crc kubenswrapper[4823]: I0320 09:45:52.558449 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d86qb" event={"ID":"7ca833ad-eed8-458f-b348-b488623a3525","Type":"ContainerStarted","Data":"fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073"} Mar 20 09:45:52 crc kubenswrapper[4823]: I0320 09:45:52.583951 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d86qb" podStartSLOduration=2.928020821 podStartE2EDuration="5.583935213s" podCreationTimestamp="2026-03-20 09:45:47 +0000 UTC" firstStartedPulling="2026-03-20 09:45:49.533519695 +0000 UTC m=+2916.219416069" lastFinishedPulling="2026-03-20 09:45:52.189434087 +0000 UTC m=+2918.875330461" observedRunningTime="2026-03-20 09:45:52.581709474 +0000 UTC m=+2919.267605868" watchObservedRunningTime="2026-03-20 09:45:52.583935213 +0000 UTC m=+2919.269831587" Mar 20 09:45:58 crc kubenswrapper[4823]: I0320 09:45:58.275542 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:58 crc kubenswrapper[4823]: I0320 09:45:58.276005 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:58 crc kubenswrapper[4823]: I0320 09:45:58.316750 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:58 crc kubenswrapper[4823]: I0320 09:45:58.648871 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:45:58 crc kubenswrapper[4823]: I0320 09:45:58.690110 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d86qb"] Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.137336 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566666-np65s"] Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.139635 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566666-np65s" Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.142135 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.142375 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.142854 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.148794 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566666-np65s"] Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.303882 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nn2s\" (UniqueName: \"kubernetes.io/projected/0bb40728-21ed-496e-ac84-a39d9adaf100-kube-api-access-4nn2s\") pod \"auto-csr-approver-29566666-np65s\" (UID: \"0bb40728-21ed-496e-ac84-a39d9adaf100\") " pod="openshift-infra/auto-csr-approver-29566666-np65s" Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.406146 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nn2s\" (UniqueName: \"kubernetes.io/projected/0bb40728-21ed-496e-ac84-a39d9adaf100-kube-api-access-4nn2s\") pod \"auto-csr-approver-29566666-np65s\" (UID: \"0bb40728-21ed-496e-ac84-a39d9adaf100\") " pod="openshift-infra/auto-csr-approver-29566666-np65s" Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.433298 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nn2s\" (UniqueName: \"kubernetes.io/projected/0bb40728-21ed-496e-ac84-a39d9adaf100-kube-api-access-4nn2s\") pod \"auto-csr-approver-29566666-np65s\" (UID: \"0bb40728-21ed-496e-ac84-a39d9adaf100\") " pod="openshift-infra/auto-csr-approver-29566666-np65s" Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.477898 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566666-np65s" Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.613859 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d86qb" podUID="7ca833ad-eed8-458f-b348-b488623a3525" containerName="registry-server" containerID="cri-o://fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073" gracePeriod=2 Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.715458 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566666-np65s"] Mar 20 09:46:00 crc kubenswrapper[4823]: I0320 09:46:00.942245 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.014803 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-utilities\") pod \"7ca833ad-eed8-458f-b348-b488623a3525\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.014913 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-catalog-content\") pod \"7ca833ad-eed8-458f-b348-b488623a3525\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.015033 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96759\" (UniqueName: \"kubernetes.io/projected/7ca833ad-eed8-458f-b348-b488623a3525-kube-api-access-96759\") pod \"7ca833ad-eed8-458f-b348-b488623a3525\" (UID: \"7ca833ad-eed8-458f-b348-b488623a3525\") " Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.015698 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-utilities" (OuterVolumeSpecName: "utilities") pod "7ca833ad-eed8-458f-b348-b488623a3525" (UID: "7ca833ad-eed8-458f-b348-b488623a3525"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.020644 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ca833ad-eed8-458f-b348-b488623a3525-kube-api-access-96759" (OuterVolumeSpecName: "kube-api-access-96759") pod "7ca833ad-eed8-458f-b348-b488623a3525" (UID: "7ca833ad-eed8-458f-b348-b488623a3525"). InnerVolumeSpecName "kube-api-access-96759". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.060396 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ca833ad-eed8-458f-b348-b488623a3525" (UID: "7ca833ad-eed8-458f-b348-b488623a3525"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.117164 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.118036 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ca833ad-eed8-458f-b348-b488623a3525-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.118065 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96759\" (UniqueName: \"kubernetes.io/projected/7ca833ad-eed8-458f-b348-b488623a3525-kube-api-access-96759\") on node \"crc\" DevicePath \"\"" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.620898 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566666-np65s" event={"ID":"0bb40728-21ed-496e-ac84-a39d9adaf100","Type":"ContainerStarted","Data":"53f52eafa9a5adbcea5569e6b7b0847657f2effeaa3c12b3ac6f78a4e543c073"} Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.623907 4823 generic.go:334] "Generic (PLEG): container finished" podID="7ca833ad-eed8-458f-b348-b488623a3525" containerID="fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073" exitCode=0 Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.623974 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d86qb" event={"ID":"7ca833ad-eed8-458f-b348-b488623a3525","Type":"ContainerDied","Data":"fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073"} Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.624022 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d86qb" event={"ID":"7ca833ad-eed8-458f-b348-b488623a3525","Type":"ContainerDied","Data":"54a9e29f295006de6ca3d0dfda1be898cb43c05fc629fbd3703a76fe2db5ca83"} Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.624059 4823 scope.go:117] "RemoveContainer" containerID="fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.624292 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d86qb" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.650831 4823 scope.go:117] "RemoveContainer" containerID="8188ec05baa2555aa0f8a3c4fb293e1cf7b8ad94d33cbe9ca2c4a1d0738e91b7" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.675844 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d86qb"] Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.681702 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d86qb"] Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.723806 4823 scope.go:117] "RemoveContainer" containerID="e1ee985454c08de0b977c6cb71ccd92810e2eb117243f3929eec4e952745314d" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.880153 4823 scope.go:117] "RemoveContainer" containerID="fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073" Mar 20 09:46:01 crc kubenswrapper[4823]: E0320 09:46:01.880969 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073\": container with ID starting with fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073 not found: ID does not exist" containerID="fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.881013 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073"} err="failed to get container status \"fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073\": rpc error: code = NotFound desc = could not find container \"fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073\": container with ID starting with fb81e98fe400493a255a954c6fa2db4ef817ac4e6d4aee87f9ee24e6f56af073 not found: ID does not exist" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.881038 4823 scope.go:117] "RemoveContainer" containerID="8188ec05baa2555aa0f8a3c4fb293e1cf7b8ad94d33cbe9ca2c4a1d0738e91b7" Mar 20 09:46:01 crc kubenswrapper[4823]: E0320 09:46:01.881310 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8188ec05baa2555aa0f8a3c4fb293e1cf7b8ad94d33cbe9ca2c4a1d0738e91b7\": container with ID starting with 8188ec05baa2555aa0f8a3c4fb293e1cf7b8ad94d33cbe9ca2c4a1d0738e91b7 not found: ID does not exist" containerID="8188ec05baa2555aa0f8a3c4fb293e1cf7b8ad94d33cbe9ca2c4a1d0738e91b7" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.881348 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8188ec05baa2555aa0f8a3c4fb293e1cf7b8ad94d33cbe9ca2c4a1d0738e91b7"} err="failed to get container status \"8188ec05baa2555aa0f8a3c4fb293e1cf7b8ad94d33cbe9ca2c4a1d0738e91b7\": rpc error: code = NotFound desc = could not find container \"8188ec05baa2555aa0f8a3c4fb293e1cf7b8ad94d33cbe9ca2c4a1d0738e91b7\": container with ID starting with 8188ec05baa2555aa0f8a3c4fb293e1cf7b8ad94d33cbe9ca2c4a1d0738e91b7 not found: ID does not exist" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.881380 4823 scope.go:117] "RemoveContainer" containerID="e1ee985454c08de0b977c6cb71ccd92810e2eb117243f3929eec4e952745314d" Mar 20 09:46:01 crc kubenswrapper[4823]: E0320 09:46:01.882252 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1ee985454c08de0b977c6cb71ccd92810e2eb117243f3929eec4e952745314d\": container with ID starting with e1ee985454c08de0b977c6cb71ccd92810e2eb117243f3929eec4e952745314d not found: ID does not exist" containerID="e1ee985454c08de0b977c6cb71ccd92810e2eb117243f3929eec4e952745314d" Mar 20 09:46:01 crc kubenswrapper[4823]: I0320 09:46:01.882277 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1ee985454c08de0b977c6cb71ccd92810e2eb117243f3929eec4e952745314d"} err="failed to get container status \"e1ee985454c08de0b977c6cb71ccd92810e2eb117243f3929eec4e952745314d\": rpc error: code = NotFound desc = could not find container \"e1ee985454c08de0b977c6cb71ccd92810e2eb117243f3929eec4e952745314d\": container with ID starting with e1ee985454c08de0b977c6cb71ccd92810e2eb117243f3929eec4e952745314d not found: ID does not exist" Mar 20 09:46:02 crc kubenswrapper[4823]: I0320 09:46:02.111819 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ca833ad-eed8-458f-b348-b488623a3525" path="/var/lib/kubelet/pods/7ca833ad-eed8-458f-b348-b488623a3525/volumes" Mar 20 09:46:02 crc kubenswrapper[4823]: I0320 09:46:02.633464 4823 generic.go:334] "Generic (PLEG): container finished" podID="0bb40728-21ed-496e-ac84-a39d9adaf100" containerID="66ee65d5905ec17f621581c57d0c5fddebd99633959722f32a839bf5ca7ded8c" exitCode=0 Mar 20 09:46:02 crc kubenswrapper[4823]: I0320 09:46:02.633529 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566666-np65s" event={"ID":"0bb40728-21ed-496e-ac84-a39d9adaf100","Type":"ContainerDied","Data":"66ee65d5905ec17f621581c57d0c5fddebd99633959722f32a839bf5ca7ded8c"} Mar 20 09:46:03 crc kubenswrapper[4823]: I0320 09:46:03.997611 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566666-np65s" Mar 20 09:46:04 crc kubenswrapper[4823]: I0320 09:46:04.085990 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nn2s\" (UniqueName: \"kubernetes.io/projected/0bb40728-21ed-496e-ac84-a39d9adaf100-kube-api-access-4nn2s\") pod \"0bb40728-21ed-496e-ac84-a39d9adaf100\" (UID: \"0bb40728-21ed-496e-ac84-a39d9adaf100\") " Mar 20 09:46:04 crc kubenswrapper[4823]: I0320 09:46:04.091361 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bb40728-21ed-496e-ac84-a39d9adaf100-kube-api-access-4nn2s" (OuterVolumeSpecName: "kube-api-access-4nn2s") pod "0bb40728-21ed-496e-ac84-a39d9adaf100" (UID: "0bb40728-21ed-496e-ac84-a39d9adaf100"). InnerVolumeSpecName "kube-api-access-4nn2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:46:04 crc kubenswrapper[4823]: I0320 09:46:04.188554 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nn2s\" (UniqueName: \"kubernetes.io/projected/0bb40728-21ed-496e-ac84-a39d9adaf100-kube-api-access-4nn2s\") on node \"crc\" DevicePath \"\"" Mar 20 09:46:04 crc kubenswrapper[4823]: I0320 09:46:04.654082 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566666-np65s" event={"ID":"0bb40728-21ed-496e-ac84-a39d9adaf100","Type":"ContainerDied","Data":"53f52eafa9a5adbcea5569e6b7b0847657f2effeaa3c12b3ac6f78a4e543c073"} Mar 20 09:46:04 crc kubenswrapper[4823]: I0320 09:46:04.654156 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53f52eafa9a5adbcea5569e6b7b0847657f2effeaa3c12b3ac6f78a4e543c073" Mar 20 09:46:04 crc kubenswrapper[4823]: I0320 09:46:04.654900 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566666-np65s" Mar 20 09:46:05 crc kubenswrapper[4823]: I0320 09:46:05.062796 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566660-vfdkw"] Mar 20 09:46:05 crc kubenswrapper[4823]: I0320 09:46:05.069357 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566660-vfdkw"] Mar 20 09:46:06 crc kubenswrapper[4823]: I0320 09:46:06.107976 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cbead88-bb8c-48a1-9e89-ef0e354cc4bb" path="/var/lib/kubelet/pods/2cbead88-bb8c-48a1-9e89-ef0e354cc4bb/volumes" Mar 20 09:46:09 crc kubenswrapper[4823]: I0320 09:46:09.430098 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:46:09 crc kubenswrapper[4823]: I0320 09:46:09.430464 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:46:39 crc kubenswrapper[4823]: I0320 09:46:39.430164 4823 patch_prober.go:28] interesting pod/machine-config-daemon-snzb2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:46:39 crc kubenswrapper[4823]: I0320 09:46:39.430769 4823 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:46:39 crc kubenswrapper[4823]: I0320 09:46:39.430835 4823 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" Mar 20 09:46:39 crc kubenswrapper[4823]: I0320 09:46:39.431569 4823 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d"} pod="openshift-machine-config-operator/machine-config-daemon-snzb2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:46:39 crc kubenswrapper[4823]: I0320 09:46:39.431660 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerName="machine-config-daemon" containerID="cri-o://1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" gracePeriod=600 Mar 20 09:46:39 crc kubenswrapper[4823]: E0320 09:46:39.561997 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:46:39 crc kubenswrapper[4823]: I0320 09:46:39.925946 4823 generic.go:334] "Generic (PLEG): container finished" podID="7322ac31-c6c4-43b9-8b3c-60ff84742545" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" exitCode=0 Mar 20 09:46:39 crc kubenswrapper[4823]: I0320 09:46:39.925992 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerDied","Data":"1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d"} Mar 20 09:46:39 crc kubenswrapper[4823]: I0320 09:46:39.926034 4823 scope.go:117] "RemoveContainer" containerID="5904f94da988d8d267e1ab7d3ea981d80fd558bb69192f0ba345d93f9d1d816c" Mar 20 09:46:39 crc kubenswrapper[4823]: I0320 09:46:39.926474 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:46:39 crc kubenswrapper[4823]: E0320 09:46:39.926692 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.453471 4823 scope.go:117] "RemoveContainer" containerID="719e58baaf61a04aa2fdb0fdd2df0380a3499ace924bf5073b6941921ed77f63" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.644789 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-r5blg/must-gather-kj82l"] Mar 20 09:46:40 crc kubenswrapper[4823]: E0320 09:46:40.645193 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ca833ad-eed8-458f-b348-b488623a3525" containerName="extract-content" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.645214 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ca833ad-eed8-458f-b348-b488623a3525" containerName="extract-content" Mar 20 09:46:40 crc kubenswrapper[4823]: E0320 09:46:40.645241 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ca833ad-eed8-458f-b348-b488623a3525" containerName="extract-utilities" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.645253 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ca833ad-eed8-458f-b348-b488623a3525" containerName="extract-utilities" Mar 20 09:46:40 crc kubenswrapper[4823]: E0320 09:46:40.645275 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ca833ad-eed8-458f-b348-b488623a3525" containerName="registry-server" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.645287 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ca833ad-eed8-458f-b348-b488623a3525" containerName="registry-server" Mar 20 09:46:40 crc kubenswrapper[4823]: E0320 09:46:40.645311 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bb40728-21ed-496e-ac84-a39d9adaf100" containerName="oc" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.645322 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bb40728-21ed-496e-ac84-a39d9adaf100" containerName="oc" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.645522 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ca833ad-eed8-458f-b348-b488623a3525" containerName="registry-server" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.645545 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bb40728-21ed-496e-ac84-a39d9adaf100" containerName="oc" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.646630 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5blg/must-gather-kj82l" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.653866 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-r5blg"/"openshift-service-ca.crt" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.658064 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-r5blg"/"default-dockercfg-scmcf" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.660807 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-r5blg"/"kube-root-ca.crt" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.668402 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-r5blg/must-gather-kj82l"] Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.757354 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/713dd23b-ced2-416e-9869-0279e6a95290-must-gather-output\") pod \"must-gather-kj82l\" (UID: \"713dd23b-ced2-416e-9869-0279e6a95290\") " pod="openshift-must-gather-r5blg/must-gather-kj82l" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.757424 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h96nv\" (UniqueName: \"kubernetes.io/projected/713dd23b-ced2-416e-9869-0279e6a95290-kube-api-access-h96nv\") pod \"must-gather-kj82l\" (UID: \"713dd23b-ced2-416e-9869-0279e6a95290\") " pod="openshift-must-gather-r5blg/must-gather-kj82l" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.859229 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/713dd23b-ced2-416e-9869-0279e6a95290-must-gather-output\") pod \"must-gather-kj82l\" (UID: \"713dd23b-ced2-416e-9869-0279e6a95290\") " pod="openshift-must-gather-r5blg/must-gather-kj82l" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.859282 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h96nv\" (UniqueName: \"kubernetes.io/projected/713dd23b-ced2-416e-9869-0279e6a95290-kube-api-access-h96nv\") pod \"must-gather-kj82l\" (UID: \"713dd23b-ced2-416e-9869-0279e6a95290\") " pod="openshift-must-gather-r5blg/must-gather-kj82l" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.859753 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/713dd23b-ced2-416e-9869-0279e6a95290-must-gather-output\") pod \"must-gather-kj82l\" (UID: \"713dd23b-ced2-416e-9869-0279e6a95290\") " pod="openshift-must-gather-r5blg/must-gather-kj82l" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.909033 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h96nv\" (UniqueName: \"kubernetes.io/projected/713dd23b-ced2-416e-9869-0279e6a95290-kube-api-access-h96nv\") pod \"must-gather-kj82l\" (UID: \"713dd23b-ced2-416e-9869-0279e6a95290\") " pod="openshift-must-gather-r5blg/must-gather-kj82l" Mar 20 09:46:40 crc kubenswrapper[4823]: I0320 09:46:40.966156 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5blg/must-gather-kj82l" Mar 20 09:46:41 crc kubenswrapper[4823]: I0320 09:46:41.398044 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-r5blg/must-gather-kj82l"] Mar 20 09:46:41 crc kubenswrapper[4823]: I0320 09:46:41.942397 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5blg/must-gather-kj82l" event={"ID":"713dd23b-ced2-416e-9869-0279e6a95290","Type":"ContainerStarted","Data":"f425ac6d747a8238a86a642f2c294799faaa92aebb638af29e81d99c11ee9fc2"} Mar 20 09:46:47 crc kubenswrapper[4823]: I0320 09:46:47.983879 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5blg/must-gather-kj82l" event={"ID":"713dd23b-ced2-416e-9869-0279e6a95290","Type":"ContainerStarted","Data":"04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37"} Mar 20 09:46:47 crc kubenswrapper[4823]: I0320 09:46:47.984339 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5blg/must-gather-kj82l" event={"ID":"713dd23b-ced2-416e-9869-0279e6a95290","Type":"ContainerStarted","Data":"9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86"} Mar 20 09:46:48 crc kubenswrapper[4823]: I0320 09:46:48.003107 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-r5blg/must-gather-kj82l" podStartSLOduration=2.596357986 podStartE2EDuration="8.003085768s" podCreationTimestamp="2026-03-20 09:46:40 +0000 UTC" firstStartedPulling="2026-03-20 09:46:41.400426657 +0000 UTC m=+2968.086323071" lastFinishedPulling="2026-03-20 09:46:46.807154479 +0000 UTC m=+2973.493050853" observedRunningTime="2026-03-20 09:46:48.000852528 +0000 UTC m=+2974.686748922" watchObservedRunningTime="2026-03-20 09:46:48.003085768 +0000 UTC m=+2974.688982152" Mar 20 09:46:52 crc kubenswrapper[4823]: I0320 09:46:52.096923 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:46:52 crc kubenswrapper[4823]: E0320 09:46:52.097553 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:47:07 crc kubenswrapper[4823]: I0320 09:47:07.097061 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:47:07 crc kubenswrapper[4823]: E0320 09:47:07.097662 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:47:19 crc kubenswrapper[4823]: I0320 09:47:19.097477 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:47:19 crc kubenswrapper[4823]: E0320 09:47:19.098449 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:47:30 crc kubenswrapper[4823]: I0320 09:47:30.059069 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dd6ddcc-t9dnq_5a2401f7-2c31-4d95-bf2c-9896d96f6ed5/init/0.log" Mar 20 09:47:30 crc kubenswrapper[4823]: I0320 09:47:30.225063 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dd6ddcc-t9dnq_5a2401f7-2c31-4d95-bf2c-9896d96f6ed5/init/0.log" Mar 20 09:47:30 crc kubenswrapper[4823]: I0320 09:47:30.259666 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dd6ddcc-t9dnq_5a2401f7-2c31-4d95-bf2c-9896d96f6ed5/dnsmasq-dns/0.log" Mar 20 09:47:33 crc kubenswrapper[4823]: I0320 09:47:33.097917 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:47:33 crc kubenswrapper[4823]: E0320 09:47:33.098647 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.743694 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4mr24"] Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.746823 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.778792 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4mr24"] Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.852694 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9518d3f7-2e03-4b2c-83a0-43f6b2401177-catalog-content\") pod \"community-operators-4mr24\" (UID: \"9518d3f7-2e03-4b2c-83a0-43f6b2401177\") " pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.853141 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9518d3f7-2e03-4b2c-83a0-43f6b2401177-utilities\") pod \"community-operators-4mr24\" (UID: \"9518d3f7-2e03-4b2c-83a0-43f6b2401177\") " pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.853178 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q6dl\" (UniqueName: \"kubernetes.io/projected/9518d3f7-2e03-4b2c-83a0-43f6b2401177-kube-api-access-2q6dl\") pod \"community-operators-4mr24\" (UID: \"9518d3f7-2e03-4b2c-83a0-43f6b2401177\") " pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.954703 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9518d3f7-2e03-4b2c-83a0-43f6b2401177-catalog-content\") pod \"community-operators-4mr24\" (UID: \"9518d3f7-2e03-4b2c-83a0-43f6b2401177\") " pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.954791 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9518d3f7-2e03-4b2c-83a0-43f6b2401177-utilities\") pod \"community-operators-4mr24\" (UID: \"9518d3f7-2e03-4b2c-83a0-43f6b2401177\") " pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.954815 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q6dl\" (UniqueName: \"kubernetes.io/projected/9518d3f7-2e03-4b2c-83a0-43f6b2401177-kube-api-access-2q6dl\") pod \"community-operators-4mr24\" (UID: \"9518d3f7-2e03-4b2c-83a0-43f6b2401177\") " pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.955290 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9518d3f7-2e03-4b2c-83a0-43f6b2401177-catalog-content\") pod \"community-operators-4mr24\" (UID: \"9518d3f7-2e03-4b2c-83a0-43f6b2401177\") " pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.955400 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9518d3f7-2e03-4b2c-83a0-43f6b2401177-utilities\") pod \"community-operators-4mr24\" (UID: \"9518d3f7-2e03-4b2c-83a0-43f6b2401177\") " pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:39 crc kubenswrapper[4823]: I0320 09:47:39.974246 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q6dl\" (UniqueName: \"kubernetes.io/projected/9518d3f7-2e03-4b2c-83a0-43f6b2401177-kube-api-access-2q6dl\") pod \"community-operators-4mr24\" (UID: \"9518d3f7-2e03-4b2c-83a0-43f6b2401177\") " pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:40 crc kubenswrapper[4823]: I0320 09:47:40.099128 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:40 crc kubenswrapper[4823]: I0320 09:47:40.575459 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4mr24"] Mar 20 09:47:41 crc kubenswrapper[4823]: I0320 09:47:41.368311 4823 generic.go:334] "Generic (PLEG): container finished" podID="9518d3f7-2e03-4b2c-83a0-43f6b2401177" containerID="f1a58516bc302905c712992adbc7fbc4cbe651c6bba2d5984f4d4df2681a4664" exitCode=0 Mar 20 09:47:41 crc kubenswrapper[4823]: I0320 09:47:41.368409 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4mr24" event={"ID":"9518d3f7-2e03-4b2c-83a0-43f6b2401177","Type":"ContainerDied","Data":"f1a58516bc302905c712992adbc7fbc4cbe651c6bba2d5984f4d4df2681a4664"} Mar 20 09:47:41 crc kubenswrapper[4823]: I0320 09:47:41.370718 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4mr24" event={"ID":"9518d3f7-2e03-4b2c-83a0-43f6b2401177","Type":"ContainerStarted","Data":"6dcb06e7f3b6d3d6fe30791808d1cf569ebacb7b45ebc480bc7cb8dc7ee196e4"} Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.136801 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jhvlg"] Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.145119 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.155383 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhvlg"] Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.293546 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbvr5\" (UniqueName: \"kubernetes.io/projected/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-kube-api-access-wbvr5\") pod \"redhat-operators-jhvlg\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.293615 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-utilities\") pod \"redhat-operators-jhvlg\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.293648 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-catalog-content\") pod \"redhat-operators-jhvlg\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.395896 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbvr5\" (UniqueName: \"kubernetes.io/projected/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-kube-api-access-wbvr5\") pod \"redhat-operators-jhvlg\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.395946 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-utilities\") pod \"redhat-operators-jhvlg\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.395967 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-catalog-content\") pod \"redhat-operators-jhvlg\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.396400 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-catalog-content\") pod \"redhat-operators-jhvlg\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.396796 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-utilities\") pod \"redhat-operators-jhvlg\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.416591 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbvr5\" (UniqueName: \"kubernetes.io/projected/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-kube-api-access-wbvr5\") pod \"redhat-operators-jhvlg\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.464232 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:42 crc kubenswrapper[4823]: I0320 09:47:42.684945 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhvlg"] Mar 20 09:47:42 crc kubenswrapper[4823]: W0320 09:47:42.695587 4823 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fd0a306_d510_4e37_a6f3_f8d3d6339f54.slice/crio-716898a4c378045d4d2b4d10ad0b95615d079e83a824cf8d704514e8d20fc005 WatchSource:0}: Error finding container 716898a4c378045d4d2b4d10ad0b95615d079e83a824cf8d704514e8d20fc005: Status 404 returned error can't find the container with id 716898a4c378045d4d2b4d10ad0b95615d079e83a824cf8d704514e8d20fc005 Mar 20 09:47:43 crc kubenswrapper[4823]: I0320 09:47:43.388707 4823 generic.go:334] "Generic (PLEG): container finished" podID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerID="11209aa25c4d49c94342fa1dd64cf0186271d2a01b29935a82758e8afaa71de0" exitCode=0 Mar 20 09:47:43 crc kubenswrapper[4823]: I0320 09:47:43.388803 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhvlg" event={"ID":"3fd0a306-d510-4e37-a6f3-f8d3d6339f54","Type":"ContainerDied","Data":"11209aa25c4d49c94342fa1dd64cf0186271d2a01b29935a82758e8afaa71de0"} Mar 20 09:47:43 crc kubenswrapper[4823]: I0320 09:47:43.388868 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhvlg" event={"ID":"3fd0a306-d510-4e37-a6f3-f8d3d6339f54","Type":"ContainerStarted","Data":"716898a4c378045d4d2b4d10ad0b95615d079e83a824cf8d704514e8d20fc005"} Mar 20 09:47:44 crc kubenswrapper[4823]: I0320 09:47:44.402955 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4_627dd666-3cdd-4c47-939c-9061526fac3d/util/0.log" Mar 20 09:47:44 crc kubenswrapper[4823]: I0320 09:47:44.403196 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhvlg" event={"ID":"3fd0a306-d510-4e37-a6f3-f8d3d6339f54","Type":"ContainerStarted","Data":"cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d"} Mar 20 09:47:44 crc kubenswrapper[4823]: I0320 09:47:44.615862 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4_627dd666-3cdd-4c47-939c-9061526fac3d/util/0.log" Mar 20 09:47:44 crc kubenswrapper[4823]: I0320 09:47:44.616503 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4_627dd666-3cdd-4c47-939c-9061526fac3d/pull/0.log" Mar 20 09:47:44 crc kubenswrapper[4823]: I0320 09:47:44.701058 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4_627dd666-3cdd-4c47-939c-9061526fac3d/pull/0.log" Mar 20 09:47:44 crc kubenswrapper[4823]: I0320 09:47:44.921047 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4_627dd666-3cdd-4c47-939c-9061526fac3d/util/0.log" Mar 20 09:47:44 crc kubenswrapper[4823]: I0320 09:47:44.939994 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4_627dd666-3cdd-4c47-939c-9061526fac3d/pull/0.log" Mar 20 09:47:44 crc kubenswrapper[4823]: I0320 09:47:44.964224 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badphzn4_627dd666-3cdd-4c47-939c-9061526fac3d/extract/0.log" Mar 20 09:47:45 crc kubenswrapper[4823]: I0320 09:47:45.081172 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59bc569d95-z9bkx_c9b3eb51-1ddc-4abc-9d8a-09611ef3983c/manager/0.log" Mar 20 09:47:45 crc kubenswrapper[4823]: I0320 09:47:45.301370 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-588d4d986b-56xqd_f99021be-4dee-4443-938e-ec54f85d61f5/manager/0.log" Mar 20 09:47:45 crc kubenswrapper[4823]: I0320 09:47:45.399712 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-79df6bcc97-7c7hb_70a65110-b953-43ca-a6ec-5131b4660aea/manager/0.log" Mar 20 09:47:45 crc kubenswrapper[4823]: I0320 09:47:45.413461 4823 generic.go:334] "Generic (PLEG): container finished" podID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerID="cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d" exitCode=0 Mar 20 09:47:45 crc kubenswrapper[4823]: I0320 09:47:45.413501 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhvlg" event={"ID":"3fd0a306-d510-4e37-a6f3-f8d3d6339f54","Type":"ContainerDied","Data":"cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d"} Mar 20 09:47:45 crc kubenswrapper[4823]: I0320 09:47:45.562452 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d58dc466-47nf2_a95e604d-a508-41bc-84e3-78fd32a90ed4/manager/0.log" Mar 20 09:47:45 crc kubenswrapper[4823]: I0320 09:47:45.594105 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-67dd5f86f5-gxz2f_8ed64ffe-e8ec-4e37-a2ba-9a7bfb86e078/manager/0.log" Mar 20 09:47:45 crc kubenswrapper[4823]: I0320 09:47:45.739265 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-8464cc45fb-75qls_ddb36285-c38c-4c75-81ca-4f55f46118e8/manager/0.log" Mar 20 09:47:45 crc kubenswrapper[4823]: I0320 09:47:45.787095 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-577ccd856-ftcjs_daa1da6a-6c56-49b4-852d-915636736f69/manager/0.log" Mar 20 09:47:45 crc kubenswrapper[4823]: I0320 09:47:45.924939 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f787dddc9-vcsmh_bf1c744e-50af-44bb-883b-085681d41cff/manager/0.log" Mar 20 09:47:45 crc kubenswrapper[4823]: I0320 09:47:45.971211 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-768b96df4c-mbb9d_cf311c3f-f3fd-47b7-9f04-427be533e8d6/manager/0.log" Mar 20 09:47:46 crc kubenswrapper[4823]: I0320 09:47:46.079214 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-55f864c847-xkcdj_d933984b-3a5c-4792-a5cb-751cae8865ef/manager/0.log" Mar 20 09:47:46 crc kubenswrapper[4823]: I0320 09:47:46.140547 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67ccfc9778-bgmxj_e19776a8-b3ac-4901-9dbf-ed46ff5b44d9/manager/0.log" Mar 20 09:47:46 crc kubenswrapper[4823]: I0320 09:47:46.302384 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-767865f676-8k6bt_570a5a06-26f8-4edb-a891-703fb297adc2/manager/0.log" Mar 20 09:47:46 crc kubenswrapper[4823]: I0320 09:47:46.343109 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5d488d59fb-28hd6_d8a80347-5759-4eb7-9750-b170fa6e5689/manager/0.log" Mar 20 09:47:46 crc kubenswrapper[4823]: I0320 09:47:46.477661 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5b9f45d989-vft4x_638e33c0-1409-4f54-a08b-25c6ef4319a7/manager/0.log" Mar 20 09:47:46 crc kubenswrapper[4823]: I0320 09:47:46.540711 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-89d64c458-rtmw4_091871f2-1d5b-4263-b190-3e6262c4d1b0/manager/0.log" Mar 20 09:47:46 crc kubenswrapper[4823]: I0320 09:47:46.829612 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-9df8dd5fd-jxnmr_328e53ac-f20f-4a71-b756-f143cd0fc077/operator/0.log" Mar 20 09:47:46 crc kubenswrapper[4823]: I0320 09:47:46.858924 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-55958644c4-vk586_dfc33271-a911-4017-bb46-3d6f5ce6ddf1/manager/0.log" Mar 20 09:47:46 crc kubenswrapper[4823]: I0320 09:47:46.931841 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-55sh4_735b38f1-4ef0-411b-97a1-f48fef038a2f/registry-server/0.log" Mar 20 09:47:47 crc kubenswrapper[4823]: I0320 09:47:47.046434 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-884679f54-fdgtg_ee7cc64f-5303-412e-83b4-322e06279025/manager/0.log" Mar 20 09:47:47 crc kubenswrapper[4823]: I0320 09:47:47.166450 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5784578c99-lp942_04565eb9-0d95-440c-a8c0-742a419cae88/manager/0.log" Mar 20 09:47:47 crc kubenswrapper[4823]: I0320 09:47:47.303338 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-c674c5965-h6dhs_98dd6787-8d0c-4612-a18c-0939eaa5a8af/manager/0.log" Mar 20 09:47:47 crc kubenswrapper[4823]: I0320 09:47:47.360228 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d6b694c5-htq9x_16e76c27-f7b9-4df7-8cc6-b1b0753613bc/manager/0.log" Mar 20 09:47:47 crc kubenswrapper[4823]: I0320 09:47:47.433361 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4mr24" event={"ID":"9518d3f7-2e03-4b2c-83a0-43f6b2401177","Type":"ContainerStarted","Data":"bf097c6e3b1e2de92d2dd72bde585d7bf095e7f1701f002acabcd3c44f1d9c37"} Mar 20 09:47:47 crc kubenswrapper[4823]: I0320 09:47:47.470501 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-8bjm7_fdc2187a-b568-473f-ab3b-0c1058789b5b/manager/0.log" Mar 20 09:47:47 crc kubenswrapper[4823]: I0320 09:47:47.544911 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-6z7z9_33c79106-0897-4b9a-ab61-5e873648f9e1/manager/0.log" Mar 20 09:47:48 crc kubenswrapper[4823]: I0320 09:47:48.097145 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:47:48 crc kubenswrapper[4823]: E0320 09:47:48.097412 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:47:48 crc kubenswrapper[4823]: I0320 09:47:48.440804 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhvlg" event={"ID":"3fd0a306-d510-4e37-a6f3-f8d3d6339f54","Type":"ContainerStarted","Data":"fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004"} Mar 20 09:47:48 crc kubenswrapper[4823]: I0320 09:47:48.443252 4823 generic.go:334] "Generic (PLEG): container finished" podID="9518d3f7-2e03-4b2c-83a0-43f6b2401177" containerID="bf097c6e3b1e2de92d2dd72bde585d7bf095e7f1701f002acabcd3c44f1d9c37" exitCode=0 Mar 20 09:47:48 crc kubenswrapper[4823]: I0320 09:47:48.443286 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4mr24" event={"ID":"9518d3f7-2e03-4b2c-83a0-43f6b2401177","Type":"ContainerDied","Data":"bf097c6e3b1e2de92d2dd72bde585d7bf095e7f1701f002acabcd3c44f1d9c37"} Mar 20 09:47:48 crc kubenswrapper[4823]: I0320 09:47:48.467557 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jhvlg" podStartSLOduration=2.4821100019999998 podStartE2EDuration="6.467538376s" podCreationTimestamp="2026-03-20 09:47:42 +0000 UTC" firstStartedPulling="2026-03-20 09:47:43.390786313 +0000 UTC m=+3030.076682687" lastFinishedPulling="2026-03-20 09:47:47.376214687 +0000 UTC m=+3034.062111061" observedRunningTime="2026-03-20 09:47:48.459264942 +0000 UTC m=+3035.145161326" watchObservedRunningTime="2026-03-20 09:47:48.467538376 +0000 UTC m=+3035.153434750" Mar 20 09:47:49 crc kubenswrapper[4823]: I0320 09:47:49.451674 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4mr24" event={"ID":"9518d3f7-2e03-4b2c-83a0-43f6b2401177","Type":"ContainerStarted","Data":"c01d89d1f90a334cf8fd30d48f2477c3348e01ed7b6fcb2b00205c34877425f7"} Mar 20 09:47:49 crc kubenswrapper[4823]: I0320 09:47:49.470169 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4mr24" podStartSLOduration=2.89274209 podStartE2EDuration="10.470148743s" podCreationTimestamp="2026-03-20 09:47:39 +0000 UTC" firstStartedPulling="2026-03-20 09:47:41.370173666 +0000 UTC m=+3028.056070040" lastFinishedPulling="2026-03-20 09:47:48.947580319 +0000 UTC m=+3035.633476693" observedRunningTime="2026-03-20 09:47:49.465887521 +0000 UTC m=+3036.151783905" watchObservedRunningTime="2026-03-20 09:47:49.470148743 +0000 UTC m=+3036.156045127" Mar 20 09:47:50 crc kubenswrapper[4823]: I0320 09:47:50.108302 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:50 crc kubenswrapper[4823]: I0320 09:47:50.108345 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:47:51 crc kubenswrapper[4823]: I0320 09:47:51.152087 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-4mr24" podUID="9518d3f7-2e03-4b2c-83a0-43f6b2401177" containerName="registry-server" probeResult="failure" output=< Mar 20 09:47:51 crc kubenswrapper[4823]: timeout: failed to connect service ":50051" within 1s Mar 20 09:47:51 crc kubenswrapper[4823]: > Mar 20 09:47:53 crc kubenswrapper[4823]: I0320 09:47:52.465114 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:53 crc kubenswrapper[4823]: I0320 09:47:52.468390 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:47:53 crc kubenswrapper[4823]: I0320 09:47:52.821949 4823 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 09:47:53 crc kubenswrapper[4823]: I0320 09:47:52.822206 4823 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 09:47:53 crc kubenswrapper[4823]: I0320 09:47:53.798678 4823 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jhvlg" podUID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerName="registry-server" probeResult="failure" output=< Mar 20 09:47:53 crc kubenswrapper[4823]: timeout: failed to connect service ":50051" within 1s Mar 20 09:47:53 crc kubenswrapper[4823]: > Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.097551 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:48:00 crc kubenswrapper[4823]: E0320 09:48:00.098236 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.135664 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566668-hdv86"] Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.136702 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566668-hdv86" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.144686 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.144728 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.144728 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.148351 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566668-hdv86"] Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.171296 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.221233 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4mr24" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.235811 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6chp\" (UniqueName: \"kubernetes.io/projected/f2910975-d36f-44de-8f46-f29434b2b208-kube-api-access-c6chp\") pod \"auto-csr-approver-29566668-hdv86\" (UID: \"f2910975-d36f-44de-8f46-f29434b2b208\") " pod="openshift-infra/auto-csr-approver-29566668-hdv86" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.298448 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4mr24"] Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.337714 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6chp\" (UniqueName: \"kubernetes.io/projected/f2910975-d36f-44de-8f46-f29434b2b208-kube-api-access-c6chp\") pod \"auto-csr-approver-29566668-hdv86\" (UID: \"f2910975-d36f-44de-8f46-f29434b2b208\") " pod="openshift-infra/auto-csr-approver-29566668-hdv86" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.355340 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6chp\" (UniqueName: \"kubernetes.io/projected/f2910975-d36f-44de-8f46-f29434b2b208-kube-api-access-c6chp\") pod \"auto-csr-approver-29566668-hdv86\" (UID: \"f2910975-d36f-44de-8f46-f29434b2b208\") " pod="openshift-infra/auto-csr-approver-29566668-hdv86" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.401079 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f5t7w"] Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.401312 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f5t7w" podUID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerName="registry-server" containerID="cri-o://23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc" gracePeriod=2 Mar 20 09:48:00 crc kubenswrapper[4823]: E0320 09:48:00.453493 4823 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc is running failed: container process not found" containerID="23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 09:48:00 crc kubenswrapper[4823]: E0320 09:48:00.454063 4823 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc is running failed: container process not found" containerID="23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 09:48:00 crc kubenswrapper[4823]: E0320 09:48:00.454309 4823 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc is running failed: container process not found" containerID="23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 09:48:00 crc kubenswrapper[4823]: E0320 09:48:00.454343 4823 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-f5t7w" podUID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerName="registry-server" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.459398 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566668-hdv86" Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.584104 4823 generic.go:334] "Generic (PLEG): container finished" podID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerID="23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc" exitCode=0 Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.584177 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5t7w" event={"ID":"f5c7111d-329b-45f4-bbd8-643fe3533f79","Type":"ContainerDied","Data":"23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc"} Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.894174 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566668-hdv86"] Mar 20 09:48:00 crc kubenswrapper[4823]: I0320 09:48:00.921149 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.051227 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-772sl\" (UniqueName: \"kubernetes.io/projected/f5c7111d-329b-45f4-bbd8-643fe3533f79-kube-api-access-772sl\") pod \"f5c7111d-329b-45f4-bbd8-643fe3533f79\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.051297 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-utilities\") pod \"f5c7111d-329b-45f4-bbd8-643fe3533f79\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.051366 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-catalog-content\") pod \"f5c7111d-329b-45f4-bbd8-643fe3533f79\" (UID: \"f5c7111d-329b-45f4-bbd8-643fe3533f79\") " Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.052689 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-utilities" (OuterVolumeSpecName: "utilities") pod "f5c7111d-329b-45f4-bbd8-643fe3533f79" (UID: "f5c7111d-329b-45f4-bbd8-643fe3533f79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.056791 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5c7111d-329b-45f4-bbd8-643fe3533f79-kube-api-access-772sl" (OuterVolumeSpecName: "kube-api-access-772sl") pod "f5c7111d-329b-45f4-bbd8-643fe3533f79" (UID: "f5c7111d-329b-45f4-bbd8-643fe3533f79"). InnerVolumeSpecName "kube-api-access-772sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.113314 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5c7111d-329b-45f4-bbd8-643fe3533f79" (UID: "f5c7111d-329b-45f4-bbd8-643fe3533f79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.153401 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-772sl\" (UniqueName: \"kubernetes.io/projected/f5c7111d-329b-45f4-bbd8-643fe3533f79-kube-api-access-772sl\") on node \"crc\" DevicePath \"\"" Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.153444 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.153456 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5c7111d-329b-45f4-bbd8-643fe3533f79-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.592133 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5t7w" Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.592054 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5t7w" event={"ID":"f5c7111d-329b-45f4-bbd8-643fe3533f79","Type":"ContainerDied","Data":"03192cd0d2e36348251c0152afe01a40bfdbb91098c60b5bebf9e42954eac82f"} Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.592564 4823 scope.go:117] "RemoveContainer" containerID="23a761f5cca1ba1bf51092fce5d6e0a127c466cb6aed19be11ab46ef8db86acc" Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.593241 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566668-hdv86" event={"ID":"f2910975-d36f-44de-8f46-f29434b2b208","Type":"ContainerStarted","Data":"8b172c94120d9b54ec65e1bb4ab5d343a6a7a1a27a22e254bc06757f4aba2b24"} Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.619668 4823 scope.go:117] "RemoveContainer" containerID="907f33c754629028b717dbbcead8f55dfe6305750aa27e1ca3804b4ce5b039a1" Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.625438 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f5t7w"] Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.636168 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f5t7w"] Mar 20 09:48:01 crc kubenswrapper[4823]: I0320 09:48:01.655821 4823 scope.go:117] "RemoveContainer" containerID="b71bc856792b5242e62166acea56c00d0f4165624bca73121fda8b33d1ad94e3" Mar 20 09:48:02 crc kubenswrapper[4823]: I0320 09:48:02.106964 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5c7111d-329b-45f4-bbd8-643fe3533f79" path="/var/lib/kubelet/pods/f5c7111d-329b-45f4-bbd8-643fe3533f79/volumes" Mar 20 09:48:02 crc kubenswrapper[4823]: I0320 09:48:02.532963 4823 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:48:02 crc kubenswrapper[4823]: I0320 09:48:02.580106 4823 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:48:02 crc kubenswrapper[4823]: I0320 09:48:02.602260 4823 generic.go:334] "Generic (PLEG): container finished" podID="f2910975-d36f-44de-8f46-f29434b2b208" containerID="926696b132d0ddd28a37be11f258c0d238343cb28256aba324b0b9bfc05460f4" exitCode=0 Mar 20 09:48:02 crc kubenswrapper[4823]: I0320 09:48:02.602301 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566668-hdv86" event={"ID":"f2910975-d36f-44de-8f46-f29434b2b208","Type":"ContainerDied","Data":"926696b132d0ddd28a37be11f258c0d238343cb28256aba324b0b9bfc05460f4"} Mar 20 09:48:03 crc kubenswrapper[4823]: I0320 09:48:03.870104 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566668-hdv86" Mar 20 09:48:03 crc kubenswrapper[4823]: I0320 09:48:03.995903 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6chp\" (UniqueName: \"kubernetes.io/projected/f2910975-d36f-44de-8f46-f29434b2b208-kube-api-access-c6chp\") pod \"f2910975-d36f-44de-8f46-f29434b2b208\" (UID: \"f2910975-d36f-44de-8f46-f29434b2b208\") " Mar 20 09:48:04 crc kubenswrapper[4823]: I0320 09:48:04.002129 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2910975-d36f-44de-8f46-f29434b2b208-kube-api-access-c6chp" (OuterVolumeSpecName: "kube-api-access-c6chp") pod "f2910975-d36f-44de-8f46-f29434b2b208" (UID: "f2910975-d36f-44de-8f46-f29434b2b208"). InnerVolumeSpecName "kube-api-access-c6chp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:48:04 crc kubenswrapper[4823]: I0320 09:48:04.097545 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6chp\" (UniqueName: \"kubernetes.io/projected/f2910975-d36f-44de-8f46-f29434b2b208-kube-api-access-c6chp\") on node \"crc\" DevicePath \"\"" Mar 20 09:48:04 crc kubenswrapper[4823]: I0320 09:48:04.601514 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jhvlg"] Mar 20 09:48:04 crc kubenswrapper[4823]: I0320 09:48:04.601781 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jhvlg" podUID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerName="registry-server" containerID="cri-o://fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004" gracePeriod=2 Mar 20 09:48:04 crc kubenswrapper[4823]: I0320 09:48:04.635071 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566668-hdv86" event={"ID":"f2910975-d36f-44de-8f46-f29434b2b208","Type":"ContainerDied","Data":"8b172c94120d9b54ec65e1bb4ab5d343a6a7a1a27a22e254bc06757f4aba2b24"} Mar 20 09:48:04 crc kubenswrapper[4823]: I0320 09:48:04.635163 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b172c94120d9b54ec65e1bb4ab5d343a6a7a1a27a22e254bc06757f4aba2b24" Mar 20 09:48:04 crc kubenswrapper[4823]: I0320 09:48:04.635383 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566668-hdv86" Mar 20 09:48:04 crc kubenswrapper[4823]: I0320 09:48:04.944398 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566662-vzc8p"] Mar 20 09:48:04 crc kubenswrapper[4823]: I0320 09:48:04.956895 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566662-vzc8p"] Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.015656 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.109890 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-catalog-content\") pod \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.109988 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbvr5\" (UniqueName: \"kubernetes.io/projected/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-kube-api-access-wbvr5\") pod \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.110032 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-utilities\") pod \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\" (UID: \"3fd0a306-d510-4e37-a6f3-f8d3d6339f54\") " Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.111175 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-utilities" (OuterVolumeSpecName: "utilities") pod "3fd0a306-d510-4e37-a6f3-f8d3d6339f54" (UID: "3fd0a306-d510-4e37-a6f3-f8d3d6339f54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.124548 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-kube-api-access-wbvr5" (OuterVolumeSpecName: "kube-api-access-wbvr5") pod "3fd0a306-d510-4e37-a6f3-f8d3d6339f54" (UID: "3fd0a306-d510-4e37-a6f3-f8d3d6339f54"). InnerVolumeSpecName "kube-api-access-wbvr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.212890 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbvr5\" (UniqueName: \"kubernetes.io/projected/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-kube-api-access-wbvr5\") on node \"crc\" DevicePath \"\"" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.213144 4823 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.237066 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3fd0a306-d510-4e37-a6f3-f8d3d6339f54" (UID: "3fd0a306-d510-4e37-a6f3-f8d3d6339f54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.314669 4823 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fd0a306-d510-4e37-a6f3-f8d3d6339f54-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.643498 4823 generic.go:334] "Generic (PLEG): container finished" podID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerID="fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004" exitCode=0 Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.643783 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhvlg" event={"ID":"3fd0a306-d510-4e37-a6f3-f8d3d6339f54","Type":"ContainerDied","Data":"fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004"} Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.643867 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhvlg" event={"ID":"3fd0a306-d510-4e37-a6f3-f8d3d6339f54","Type":"ContainerDied","Data":"716898a4c378045d4d2b4d10ad0b95615d079e83a824cf8d704514e8d20fc005"} Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.643934 4823 scope.go:117] "RemoveContainer" containerID="fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.644102 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhvlg" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.672153 4823 scope.go:117] "RemoveContainer" containerID="cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.673885 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jhvlg"] Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.681098 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jhvlg"] Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.692133 4823 scope.go:117] "RemoveContainer" containerID="11209aa25c4d49c94342fa1dd64cf0186271d2a01b29935a82758e8afaa71de0" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.722569 4823 scope.go:117] "RemoveContainer" containerID="fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004" Mar 20 09:48:05 crc kubenswrapper[4823]: E0320 09:48:05.723143 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004\": container with ID starting with fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004 not found: ID does not exist" containerID="fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.723191 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004"} err="failed to get container status \"fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004\": rpc error: code = NotFound desc = could not find container \"fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004\": container with ID starting with fa2004ae3554d96173c8bbbe826cc6329ae487f2110a0f26e55304c32f11e004 not found: ID does not exist" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.723222 4823 scope.go:117] "RemoveContainer" containerID="cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d" Mar 20 09:48:05 crc kubenswrapper[4823]: E0320 09:48:05.723653 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d\": container with ID starting with cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d not found: ID does not exist" containerID="cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.723683 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d"} err="failed to get container status \"cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d\": rpc error: code = NotFound desc = could not find container \"cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d\": container with ID starting with cb653ba3e03ad943b13231297d8b77126520e67d4e2308d51e2b621c837b874d not found: ID does not exist" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.723706 4823 scope.go:117] "RemoveContainer" containerID="11209aa25c4d49c94342fa1dd64cf0186271d2a01b29935a82758e8afaa71de0" Mar 20 09:48:05 crc kubenswrapper[4823]: E0320 09:48:05.724020 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11209aa25c4d49c94342fa1dd64cf0186271d2a01b29935a82758e8afaa71de0\": container with ID starting with 11209aa25c4d49c94342fa1dd64cf0186271d2a01b29935a82758e8afaa71de0 not found: ID does not exist" containerID="11209aa25c4d49c94342fa1dd64cf0186271d2a01b29935a82758e8afaa71de0" Mar 20 09:48:05 crc kubenswrapper[4823]: I0320 09:48:05.724059 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11209aa25c4d49c94342fa1dd64cf0186271d2a01b29935a82758e8afaa71de0"} err="failed to get container status \"11209aa25c4d49c94342fa1dd64cf0186271d2a01b29935a82758e8afaa71de0\": rpc error: code = NotFound desc = could not find container \"11209aa25c4d49c94342fa1dd64cf0186271d2a01b29935a82758e8afaa71de0\": container with ID starting with 11209aa25c4d49c94342fa1dd64cf0186271d2a01b29935a82758e8afaa71de0 not found: ID does not exist" Mar 20 09:48:06 crc kubenswrapper[4823]: I0320 09:48:06.106328 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15647a49-cbe7-4a32-b42c-60c1fbb0b0ca" path="/var/lib/kubelet/pods/15647a49-cbe7-4a32-b42c-60c1fbb0b0ca/volumes" Mar 20 09:48:06 crc kubenswrapper[4823]: I0320 09:48:06.107066 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" path="/var/lib/kubelet/pods/3fd0a306-d510-4e37-a6f3-f8d3d6339f54/volumes" Mar 20 09:48:08 crc kubenswrapper[4823]: I0320 09:48:08.933442 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qst64_4da55054-ed9d-4d06-a94c-5f225541f7ea/control-plane-machine-set-operator/0.log" Mar 20 09:48:09 crc kubenswrapper[4823]: I0320 09:48:09.116245 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-49lkt_c8afe39d-c5e2-4241-a8ac-64f32c8c31fe/kube-rbac-proxy/0.log" Mar 20 09:48:09 crc kubenswrapper[4823]: I0320 09:48:09.125659 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-49lkt_c8afe39d-c5e2-4241-a8ac-64f32c8c31fe/machine-api-operator/0.log" Mar 20 09:48:14 crc kubenswrapper[4823]: I0320 09:48:14.113421 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:48:14 crc kubenswrapper[4823]: E0320 09:48:14.114636 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:48:21 crc kubenswrapper[4823]: I0320 09:48:21.603454 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-nkqmf_db934289-d3d9-4319-a56f-37faa9181e58/cert-manager-controller/0.log" Mar 20 09:48:21 crc kubenswrapper[4823]: I0320 09:48:21.786409 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-hfnfq_393e0c29-456f-4547-a8fd-7f14e3b5285c/cert-manager-cainjector/0.log" Mar 20 09:48:21 crc kubenswrapper[4823]: I0320 09:48:21.800991 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-8jmq6_d65f0664-ecd0-4123-81e3-39ec448311d4/cert-manager-webhook/0.log" Mar 20 09:48:25 crc kubenswrapper[4823]: I0320 09:48:25.097792 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:48:25 crc kubenswrapper[4823]: E0320 09:48:25.098567 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:48:33 crc kubenswrapper[4823]: I0320 09:48:33.099265 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-58bd4_23d06143-924e-49a4-bfcd-d1e29cbd247a/nmstate-console-plugin/0.log" Mar 20 09:48:33 crc kubenswrapper[4823]: I0320 09:48:33.275814 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-rm282_76f878ff-14b3-45f2-9c2b-6c4995ab52b2/nmstate-handler/0.log" Mar 20 09:48:33 crc kubenswrapper[4823]: I0320 09:48:33.376317 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-lm2pw_8af45186-c303-4a66-ad3a-bd1d59df752e/kube-rbac-proxy/0.log" Mar 20 09:48:33 crc kubenswrapper[4823]: I0320 09:48:33.385508 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-lm2pw_8af45186-c303-4a66-ad3a-bd1d59df752e/nmstate-metrics/0.log" Mar 20 09:48:33 crc kubenswrapper[4823]: I0320 09:48:33.510414 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-mfn9c_b5f0a310-2e6a-4ed5-8703-032996455052/nmstate-operator/0.log" Mar 20 09:48:33 crc kubenswrapper[4823]: I0320 09:48:33.603471 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-qfwd5_a7ccbb68-47d3-4f1b-8836-fb9f820b9b8e/nmstate-webhook/0.log" Mar 20 09:48:38 crc kubenswrapper[4823]: I0320 09:48:38.097499 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:48:38 crc kubenswrapper[4823]: E0320 09:48:38.098053 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:48:40 crc kubenswrapper[4823]: I0320 09:48:40.582148 4823 scope.go:117] "RemoveContainer" containerID="ed7feeb67421feb80da3e759844f08c6a7a7abf8e357fb567a029e4fd299420f" Mar 20 09:48:53 crc kubenswrapper[4823]: I0320 09:48:53.097780 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:48:53 crc kubenswrapper[4823]: E0320 09:48:53.098685 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:48:58 crc kubenswrapper[4823]: I0320 09:48:58.768940 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-jh928_1de94133-3c13-40f1-980f-da1342011a04/kube-rbac-proxy/0.log" Mar 20 09:48:58 crc kubenswrapper[4823]: I0320 09:48:58.829259 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-jh928_1de94133-3c13-40f1-980f-da1342011a04/controller/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.122577 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-trvvd_ef072780-7e18-4446-b701-501534f52d8d/frr-k8s-webhook-server/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.140766 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-frr-files/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.262799 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-reloader/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.287561 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-metrics/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.343552 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-frr-files/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.354371 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-reloader/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.471274 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-frr-files/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.497736 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-reloader/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.537894 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-metrics/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.599273 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-metrics/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.724963 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-reloader/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.740705 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-metrics/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.802861 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/controller/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.803320 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/cp-frr-files/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.890240 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/frr-metrics/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.970447 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/kube-rbac-proxy/0.log" Mar 20 09:48:59 crc kubenswrapper[4823]: I0320 09:48:59.999254 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/kube-rbac-proxy-frr/0.log" Mar 20 09:49:00 crc kubenswrapper[4823]: I0320 09:49:00.110832 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/reloader/0.log" Mar 20 09:49:00 crc kubenswrapper[4823]: I0320 09:49:00.251834 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-855dc6dbf9-tw2n2_d4d5d92a-98ad-40ff-9b4e-4568375f78b4/manager/0.log" Mar 20 09:49:00 crc kubenswrapper[4823]: I0320 09:49:00.394962 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5d4db96fb4-qjtz4_f6438eac-8065-43a7-9c1a-eabc0a7ac277/webhook-server/0.log" Mar 20 09:49:00 crc kubenswrapper[4823]: I0320 09:49:00.436692 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zgtrw_00eeb55c-4fd8-4aa3-ac81-b687c2074a5e/frr/0.log" Mar 20 09:49:00 crc kubenswrapper[4823]: I0320 09:49:00.531712 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cmc9m_ef19c0e0-efc5-4902-beb9-99b865a664c1/kube-rbac-proxy/0.log" Mar 20 09:49:00 crc kubenswrapper[4823]: I0320 09:49:00.690308 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cmc9m_ef19c0e0-efc5-4902-beb9-99b865a664c1/speaker/0.log" Mar 20 09:49:08 crc kubenswrapper[4823]: I0320 09:49:08.099541 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:49:08 crc kubenswrapper[4823]: E0320 09:49:08.100466 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:49:13 crc kubenswrapper[4823]: I0320 09:49:13.080107 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4_66e1071b-3fa5-454b-aca8-ec2fa91d195e/util/0.log" Mar 20 09:49:13 crc kubenswrapper[4823]: I0320 09:49:13.283117 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4_66e1071b-3fa5-454b-aca8-ec2fa91d195e/pull/0.log" Mar 20 09:49:13 crc kubenswrapper[4823]: I0320 09:49:13.290591 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4_66e1071b-3fa5-454b-aca8-ec2fa91d195e/util/0.log" Mar 20 09:49:13 crc kubenswrapper[4823]: I0320 09:49:13.359657 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4_66e1071b-3fa5-454b-aca8-ec2fa91d195e/pull/0.log" Mar 20 09:49:13 crc kubenswrapper[4823]: I0320 09:49:13.487066 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4_66e1071b-3fa5-454b-aca8-ec2fa91d195e/util/0.log" Mar 20 09:49:13 crc kubenswrapper[4823]: I0320 09:49:13.537762 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4_66e1071b-3fa5-454b-aca8-ec2fa91d195e/extract/0.log" Mar 20 09:49:13 crc kubenswrapper[4823]: I0320 09:49:13.538480 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874bb7h4_66e1071b-3fa5-454b-aca8-ec2fa91d195e/pull/0.log" Mar 20 09:49:13 crc kubenswrapper[4823]: I0320 09:49:13.682721 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc_33cd5d1c-5e85-4f7b-a322-b78cf13812d5/util/0.log" Mar 20 09:49:13 crc kubenswrapper[4823]: I0320 09:49:13.824072 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc_33cd5d1c-5e85-4f7b-a322-b78cf13812d5/pull/0.log" Mar 20 09:49:13 crc kubenswrapper[4823]: I0320 09:49:13.849873 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc_33cd5d1c-5e85-4f7b-a322-b78cf13812d5/pull/0.log" Mar 20 09:49:13 crc kubenswrapper[4823]: I0320 09:49:13.868562 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc_33cd5d1c-5e85-4f7b-a322-b78cf13812d5/util/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.004263 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc_33cd5d1c-5e85-4f7b-a322-b78cf13812d5/pull/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.039121 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc_33cd5d1c-5e85-4f7b-a322-b78cf13812d5/extract/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.052428 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c18svbc_33cd5d1c-5e85-4f7b-a322-b78cf13812d5/util/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.183606 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2gqwq_1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162/extract-utilities/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.361849 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2gqwq_1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162/extract-utilities/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.363916 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2gqwq_1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162/extract-content/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.390552 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2gqwq_1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162/extract-content/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.569271 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2gqwq_1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162/extract-utilities/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.573964 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2gqwq_1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162/extract-content/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.829789 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4mr24_9518d3f7-2e03-4b2c-83a0-43f6b2401177/extract-utilities/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.951784 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4mr24_9518d3f7-2e03-4b2c-83a0-43f6b2401177/extract-utilities/0.log" Mar 20 09:49:14 crc kubenswrapper[4823]: I0320 09:49:14.989237 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4mr24_9518d3f7-2e03-4b2c-83a0-43f6b2401177/extract-content/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.056617 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4mr24_9518d3f7-2e03-4b2c-83a0-43f6b2401177/extract-content/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.070105 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2gqwq_1b0b7f3c-8636-4aa8-a7a2-0a7aa6c75162/registry-server/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.220810 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4mr24_9518d3f7-2e03-4b2c-83a0-43f6b2401177/extract-utilities/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.239911 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4mr24_9518d3f7-2e03-4b2c-83a0-43f6b2401177/extract-content/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.356757 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4mr24_9518d3f7-2e03-4b2c-83a0-43f6b2401177/registry-server/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.435528 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qcvh5_a5f8c83b-29f6-4dca-bee6-7bc332986168/marketplace-operator/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.531327 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j8d5w_acd3ec8c-1372-46ef-9c71-9c6185bfd892/extract-utilities/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.713831 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j8d5w_acd3ec8c-1372-46ef-9c71-9c6185bfd892/extract-content/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.723582 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j8d5w_acd3ec8c-1372-46ef-9c71-9c6185bfd892/extract-content/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.738683 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j8d5w_acd3ec8c-1372-46ef-9c71-9c6185bfd892/extract-utilities/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.904501 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j8d5w_acd3ec8c-1372-46ef-9c71-9c6185bfd892/extract-utilities/0.log" Mar 20 09:49:15 crc kubenswrapper[4823]: I0320 09:49:15.922222 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j8d5w_acd3ec8c-1372-46ef-9c71-9c6185bfd892/extract-content/0.log" Mar 20 09:49:16 crc kubenswrapper[4823]: I0320 09:49:16.053145 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j8d5w_acd3ec8c-1372-46ef-9c71-9c6185bfd892/registry-server/0.log" Mar 20 09:49:16 crc kubenswrapper[4823]: I0320 09:49:16.107316 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wzt57_0a6b6c50-171e-4eec-acf0-29dc3406e657/extract-utilities/0.log" Mar 20 09:49:16 crc kubenswrapper[4823]: I0320 09:49:16.273771 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wzt57_0a6b6c50-171e-4eec-acf0-29dc3406e657/extract-utilities/0.log" Mar 20 09:49:16 crc kubenswrapper[4823]: I0320 09:49:16.300409 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wzt57_0a6b6c50-171e-4eec-acf0-29dc3406e657/extract-content/0.log" Mar 20 09:49:16 crc kubenswrapper[4823]: I0320 09:49:16.317546 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wzt57_0a6b6c50-171e-4eec-acf0-29dc3406e657/extract-content/0.log" Mar 20 09:49:16 crc kubenswrapper[4823]: I0320 09:49:16.445579 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wzt57_0a6b6c50-171e-4eec-acf0-29dc3406e657/extract-utilities/0.log" Mar 20 09:49:16 crc kubenswrapper[4823]: I0320 09:49:16.486921 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wzt57_0a6b6c50-171e-4eec-acf0-29dc3406e657/extract-content/0.log" Mar 20 09:49:16 crc kubenswrapper[4823]: I0320 09:49:16.953316 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wzt57_0a6b6c50-171e-4eec-acf0-29dc3406e657/registry-server/0.log" Mar 20 09:49:23 crc kubenswrapper[4823]: I0320 09:49:23.097558 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:49:23 crc kubenswrapper[4823]: E0320 09:49:23.098537 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:49:37 crc kubenswrapper[4823]: I0320 09:49:37.097595 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:49:37 crc kubenswrapper[4823]: E0320 09:49:37.098628 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:49:52 crc kubenswrapper[4823]: I0320 09:49:52.097442 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:49:52 crc kubenswrapper[4823]: E0320 09:49:52.098281 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.162317 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566670-k4x2g"] Mar 20 09:50:00 crc kubenswrapper[4823]: E0320 09:50:00.163381 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerName="extract-content" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.163404 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerName="extract-content" Mar 20 09:50:00 crc kubenswrapper[4823]: E0320 09:50:00.163418 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerName="registry-server" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.163429 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerName="registry-server" Mar 20 09:50:00 crc kubenswrapper[4823]: E0320 09:50:00.163442 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerName="extract-content" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.163454 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerName="extract-content" Mar 20 09:50:00 crc kubenswrapper[4823]: E0320 09:50:00.163485 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerName="registry-server" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.163496 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerName="registry-server" Mar 20 09:50:00 crc kubenswrapper[4823]: E0320 09:50:00.163517 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerName="extract-utilities" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.163527 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerName="extract-utilities" Mar 20 09:50:00 crc kubenswrapper[4823]: E0320 09:50:00.163550 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2910975-d36f-44de-8f46-f29434b2b208" containerName="oc" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.163561 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2910975-d36f-44de-8f46-f29434b2b208" containerName="oc" Mar 20 09:50:00 crc kubenswrapper[4823]: E0320 09:50:00.163579 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerName="extract-utilities" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.163590 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerName="extract-utilities" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.163834 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5c7111d-329b-45f4-bbd8-643fe3533f79" containerName="registry-server" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.163854 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fd0a306-d510-4e37-a6f3-f8d3d6339f54" containerName="registry-server" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.163880 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2910975-d36f-44de-8f46-f29434b2b208" containerName="oc" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.164522 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566670-k4x2g" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.168721 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.168721 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.169194 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.172541 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566670-k4x2g"] Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.213970 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zqch\" (UniqueName: \"kubernetes.io/projected/2675cbca-b17f-4347-892b-8df35324254a-kube-api-access-8zqch\") pod \"auto-csr-approver-29566670-k4x2g\" (UID: \"2675cbca-b17f-4347-892b-8df35324254a\") " pod="openshift-infra/auto-csr-approver-29566670-k4x2g" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.314958 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zqch\" (UniqueName: \"kubernetes.io/projected/2675cbca-b17f-4347-892b-8df35324254a-kube-api-access-8zqch\") pod \"auto-csr-approver-29566670-k4x2g\" (UID: \"2675cbca-b17f-4347-892b-8df35324254a\") " pod="openshift-infra/auto-csr-approver-29566670-k4x2g" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.340327 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zqch\" (UniqueName: \"kubernetes.io/projected/2675cbca-b17f-4347-892b-8df35324254a-kube-api-access-8zqch\") pod \"auto-csr-approver-29566670-k4x2g\" (UID: \"2675cbca-b17f-4347-892b-8df35324254a\") " pod="openshift-infra/auto-csr-approver-29566670-k4x2g" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.490480 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566670-k4x2g" Mar 20 09:50:00 crc kubenswrapper[4823]: I0320 09:50:00.969368 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566670-k4x2g"] Mar 20 09:50:01 crc kubenswrapper[4823]: I0320 09:50:01.546663 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566670-k4x2g" event={"ID":"2675cbca-b17f-4347-892b-8df35324254a","Type":"ContainerStarted","Data":"c373b6a712231e729e95b0bae580dad758d67a8daadd34bfb494e6bb487ec05d"} Mar 20 09:50:02 crc kubenswrapper[4823]: I0320 09:50:02.553918 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566670-k4x2g" event={"ID":"2675cbca-b17f-4347-892b-8df35324254a","Type":"ContainerStarted","Data":"1d1915f3468ce842807c9494cd660e97f2b198f6db43001331a3907c4dfb98ca"} Mar 20 09:50:02 crc kubenswrapper[4823]: I0320 09:50:02.572935 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566670-k4x2g" podStartSLOduration=1.315854933 podStartE2EDuration="2.572911785s" podCreationTimestamp="2026-03-20 09:50:00 +0000 UTC" firstStartedPulling="2026-03-20 09:50:00.96962018 +0000 UTC m=+3167.655516554" lastFinishedPulling="2026-03-20 09:50:02.226677032 +0000 UTC m=+3168.912573406" observedRunningTime="2026-03-20 09:50:02.566701554 +0000 UTC m=+3169.252597928" watchObservedRunningTime="2026-03-20 09:50:02.572911785 +0000 UTC m=+3169.258808179" Mar 20 09:50:03 crc kubenswrapper[4823]: I0320 09:50:03.566597 4823 generic.go:334] "Generic (PLEG): container finished" podID="2675cbca-b17f-4347-892b-8df35324254a" containerID="1d1915f3468ce842807c9494cd660e97f2b198f6db43001331a3907c4dfb98ca" exitCode=0 Mar 20 09:50:03 crc kubenswrapper[4823]: I0320 09:50:03.566843 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566670-k4x2g" event={"ID":"2675cbca-b17f-4347-892b-8df35324254a","Type":"ContainerDied","Data":"1d1915f3468ce842807c9494cd660e97f2b198f6db43001331a3907c4dfb98ca"} Mar 20 09:50:04 crc kubenswrapper[4823]: I0320 09:50:04.842938 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566670-k4x2g" Mar 20 09:50:04 crc kubenswrapper[4823]: I0320 09:50:04.986649 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zqch\" (UniqueName: \"kubernetes.io/projected/2675cbca-b17f-4347-892b-8df35324254a-kube-api-access-8zqch\") pod \"2675cbca-b17f-4347-892b-8df35324254a\" (UID: \"2675cbca-b17f-4347-892b-8df35324254a\") " Mar 20 09:50:04 crc kubenswrapper[4823]: I0320 09:50:04.992662 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2675cbca-b17f-4347-892b-8df35324254a-kube-api-access-8zqch" (OuterVolumeSpecName: "kube-api-access-8zqch") pod "2675cbca-b17f-4347-892b-8df35324254a" (UID: "2675cbca-b17f-4347-892b-8df35324254a"). InnerVolumeSpecName "kube-api-access-8zqch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:50:05 crc kubenswrapper[4823]: I0320 09:50:05.088160 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zqch\" (UniqueName: \"kubernetes.io/projected/2675cbca-b17f-4347-892b-8df35324254a-kube-api-access-8zqch\") on node \"crc\" DevicePath \"\"" Mar 20 09:50:05 crc kubenswrapper[4823]: I0320 09:50:05.097889 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:50:05 crc kubenswrapper[4823]: E0320 09:50:05.098083 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:50:05 crc kubenswrapper[4823]: I0320 09:50:05.591355 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566670-k4x2g" event={"ID":"2675cbca-b17f-4347-892b-8df35324254a","Type":"ContainerDied","Data":"c373b6a712231e729e95b0bae580dad758d67a8daadd34bfb494e6bb487ec05d"} Mar 20 09:50:05 crc kubenswrapper[4823]: I0320 09:50:05.591399 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566670-k4x2g" Mar 20 09:50:05 crc kubenswrapper[4823]: I0320 09:50:05.591398 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c373b6a712231e729e95b0bae580dad758d67a8daadd34bfb494e6bb487ec05d" Mar 20 09:50:05 crc kubenswrapper[4823]: I0320 09:50:05.660632 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566664-cb7px"] Mar 20 09:50:05 crc kubenswrapper[4823]: I0320 09:50:05.665716 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566664-cb7px"] Mar 20 09:50:05 crc kubenswrapper[4823]: E0320 09:50:05.685859 4823 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2675cbca_b17f_4347_892b_8df35324254a.slice\": RecentStats: unable to find data in memory cache]" Mar 20 09:50:06 crc kubenswrapper[4823]: I0320 09:50:06.117289 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eff8c9ec-4a29-4fdb-bb9f-375e6d48151d" path="/var/lib/kubelet/pods/eff8c9ec-4a29-4fdb-bb9f-375e6d48151d/volumes" Mar 20 09:50:16 crc kubenswrapper[4823]: I0320 09:50:16.101104 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:50:16 crc kubenswrapper[4823]: E0320 09:50:16.101723 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:50:26 crc kubenswrapper[4823]: I0320 09:50:26.733633 4823 generic.go:334] "Generic (PLEG): container finished" podID="713dd23b-ced2-416e-9869-0279e6a95290" containerID="9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86" exitCode=0 Mar 20 09:50:26 crc kubenswrapper[4823]: I0320 09:50:26.733816 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5blg/must-gather-kj82l" event={"ID":"713dd23b-ced2-416e-9869-0279e6a95290","Type":"ContainerDied","Data":"9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86"} Mar 20 09:50:26 crc kubenswrapper[4823]: I0320 09:50:26.735070 4823 scope.go:117] "RemoveContainer" containerID="9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86" Mar 20 09:50:27 crc kubenswrapper[4823]: I0320 09:50:27.096620 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:50:27 crc kubenswrapper[4823]: E0320 09:50:27.096853 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:50:27 crc kubenswrapper[4823]: I0320 09:50:27.174454 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-r5blg_must-gather-kj82l_713dd23b-ced2-416e-9869-0279e6a95290/gather/0.log" Mar 20 09:50:33 crc kubenswrapper[4823]: I0320 09:50:33.944584 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-r5blg/must-gather-kj82l"] Mar 20 09:50:33 crc kubenswrapper[4823]: I0320 09:50:33.945299 4823 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-r5blg/must-gather-kj82l" podUID="713dd23b-ced2-416e-9869-0279e6a95290" containerName="copy" containerID="cri-o://04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37" gracePeriod=2 Mar 20 09:50:33 crc kubenswrapper[4823]: I0320 09:50:33.949785 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-r5blg/must-gather-kj82l"] Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.361187 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-r5blg_must-gather-kj82l_713dd23b-ced2-416e-9869-0279e6a95290/copy/0.log" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.362194 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5blg/must-gather-kj82l" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.440835 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/713dd23b-ced2-416e-9869-0279e6a95290-must-gather-output\") pod \"713dd23b-ced2-416e-9869-0279e6a95290\" (UID: \"713dd23b-ced2-416e-9869-0279e6a95290\") " Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.441026 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h96nv\" (UniqueName: \"kubernetes.io/projected/713dd23b-ced2-416e-9869-0279e6a95290-kube-api-access-h96nv\") pod \"713dd23b-ced2-416e-9869-0279e6a95290\" (UID: \"713dd23b-ced2-416e-9869-0279e6a95290\") " Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.446154 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/713dd23b-ced2-416e-9869-0279e6a95290-kube-api-access-h96nv" (OuterVolumeSpecName: "kube-api-access-h96nv") pod "713dd23b-ced2-416e-9869-0279e6a95290" (UID: "713dd23b-ced2-416e-9869-0279e6a95290"). InnerVolumeSpecName "kube-api-access-h96nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.530172 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/713dd23b-ced2-416e-9869-0279e6a95290-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "713dd23b-ced2-416e-9869-0279e6a95290" (UID: "713dd23b-ced2-416e-9869-0279e6a95290"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.542466 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h96nv\" (UniqueName: \"kubernetes.io/projected/713dd23b-ced2-416e-9869-0279e6a95290-kube-api-access-h96nv\") on node \"crc\" DevicePath \"\"" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.542497 4823 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/713dd23b-ced2-416e-9869-0279e6a95290-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.789956 4823 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-r5blg_must-gather-kj82l_713dd23b-ced2-416e-9869-0279e6a95290/copy/0.log" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.790869 4823 generic.go:334] "Generic (PLEG): container finished" podID="713dd23b-ced2-416e-9869-0279e6a95290" containerID="04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37" exitCode=143 Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.790945 4823 scope.go:117] "RemoveContainer" containerID="04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.790976 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5blg/must-gather-kj82l" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.823590 4823 scope.go:117] "RemoveContainer" containerID="9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.882226 4823 scope.go:117] "RemoveContainer" containerID="04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37" Mar 20 09:50:34 crc kubenswrapper[4823]: E0320 09:50:34.882728 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37\": container with ID starting with 04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37 not found: ID does not exist" containerID="04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.882777 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37"} err="failed to get container status \"04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37\": rpc error: code = NotFound desc = could not find container \"04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37\": container with ID starting with 04de0ef2bc8d5c5d0b301a3afa5aafc79f4f1be22184c319509eddbab8174f37 not found: ID does not exist" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.882803 4823 scope.go:117] "RemoveContainer" containerID="9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86" Mar 20 09:50:34 crc kubenswrapper[4823]: E0320 09:50:34.883126 4823 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86\": container with ID starting with 9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86 not found: ID does not exist" containerID="9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86" Mar 20 09:50:34 crc kubenswrapper[4823]: I0320 09:50:34.883169 4823 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86"} err="failed to get container status \"9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86\": rpc error: code = NotFound desc = could not find container \"9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86\": container with ID starting with 9d9a5de100b1503735b43969199ba1674c8986546b9760a15576c98219323b86 not found: ID does not exist" Mar 20 09:50:36 crc kubenswrapper[4823]: I0320 09:50:36.112939 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="713dd23b-ced2-416e-9869-0279e6a95290" path="/var/lib/kubelet/pods/713dd23b-ced2-416e-9869-0279e6a95290/volumes" Mar 20 09:50:40 crc kubenswrapper[4823]: I0320 09:50:40.700762 4823 scope.go:117] "RemoveContainer" containerID="5254cdd127841f3ab42734cd7833c644b1cdf0e81541365cac2f53ac48b9c09e" Mar 20 09:50:41 crc kubenswrapper[4823]: I0320 09:50:41.097898 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:50:41 crc kubenswrapper[4823]: E0320 09:50:41.098283 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:50:56 crc kubenswrapper[4823]: I0320 09:50:56.098038 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:50:56 crc kubenswrapper[4823]: E0320 09:50:56.099113 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:51:11 crc kubenswrapper[4823]: I0320 09:51:11.097805 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:51:11 crc kubenswrapper[4823]: E0320 09:51:11.098918 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:51:23 crc kubenswrapper[4823]: I0320 09:51:23.097527 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:51:23 crc kubenswrapper[4823]: E0320 09:51:23.098460 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:51:37 crc kubenswrapper[4823]: I0320 09:51:37.096990 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:51:37 crc kubenswrapper[4823]: E0320 09:51:37.097651 4823 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-snzb2_openshift-machine-config-operator(7322ac31-c6c4-43b9-8b3c-60ff84742545)\"" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" podUID="7322ac31-c6c4-43b9-8b3c-60ff84742545" Mar 20 09:51:50 crc kubenswrapper[4823]: I0320 09:51:50.097390 4823 scope.go:117] "RemoveContainer" containerID="1b5ebcce419ed335f056bf48975e578be1ad9f0fb7d6e20907a17eddf46d416d" Mar 20 09:51:50 crc kubenswrapper[4823]: I0320 09:51:50.410983 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-snzb2" event={"ID":"7322ac31-c6c4-43b9-8b3c-60ff84742545","Type":"ContainerStarted","Data":"b8135cd5c5374583e5b6dc79b13cb5733dc3b1c79aa83f30e87915dbfbc7adc5"} Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.188794 4823 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566672-k6x6x"] Mar 20 09:52:00 crc kubenswrapper[4823]: E0320 09:52:00.189684 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="713dd23b-ced2-416e-9869-0279e6a95290" containerName="copy" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.189700 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="713dd23b-ced2-416e-9869-0279e6a95290" containerName="copy" Mar 20 09:52:00 crc kubenswrapper[4823]: E0320 09:52:00.189720 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2675cbca-b17f-4347-892b-8df35324254a" containerName="oc" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.189728 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="2675cbca-b17f-4347-892b-8df35324254a" containerName="oc" Mar 20 09:52:00 crc kubenswrapper[4823]: E0320 09:52:00.189771 4823 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="713dd23b-ced2-416e-9869-0279e6a95290" containerName="gather" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.189781 4823 state_mem.go:107] "Deleted CPUSet assignment" podUID="713dd23b-ced2-416e-9869-0279e6a95290" containerName="gather" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.189938 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="713dd23b-ced2-416e-9869-0279e6a95290" containerName="gather" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.189956 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="713dd23b-ced2-416e-9869-0279e6a95290" containerName="copy" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.189970 4823 memory_manager.go:354] "RemoveStaleState removing state" podUID="2675cbca-b17f-4347-892b-8df35324254a" containerName="oc" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.190478 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566672-k6x6x" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.192980 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.194002 4823 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.194606 4823 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-zcc8r" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.195901 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566672-k6x6x"] Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.376603 4823 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9qbc\" (UniqueName: \"kubernetes.io/projected/09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d-kube-api-access-q9qbc\") pod \"auto-csr-approver-29566672-k6x6x\" (UID: \"09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d\") " pod="openshift-infra/auto-csr-approver-29566672-k6x6x" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.478768 4823 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9qbc\" (UniqueName: \"kubernetes.io/projected/09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d-kube-api-access-q9qbc\") pod \"auto-csr-approver-29566672-k6x6x\" (UID: \"09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d\") " pod="openshift-infra/auto-csr-approver-29566672-k6x6x" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.515119 4823 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9qbc\" (UniqueName: \"kubernetes.io/projected/09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d-kube-api-access-q9qbc\") pod \"auto-csr-approver-29566672-k6x6x\" (UID: \"09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d\") " pod="openshift-infra/auto-csr-approver-29566672-k6x6x" Mar 20 09:52:00 crc kubenswrapper[4823]: I0320 09:52:00.807319 4823 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566672-k6x6x" Mar 20 09:52:01 crc kubenswrapper[4823]: I0320 09:52:01.241604 4823 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566672-k6x6x"] Mar 20 09:52:01 crc kubenswrapper[4823]: I0320 09:52:01.251495 4823 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:52:01 crc kubenswrapper[4823]: I0320 09:52:01.495558 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566672-k6x6x" event={"ID":"09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d","Type":"ContainerStarted","Data":"57abb0b7f873fc862a148ca4409f89fca8f5a3d5c69c6e5527db83fc5c0b0674"} Mar 20 09:52:02 crc kubenswrapper[4823]: I0320 09:52:02.508118 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566672-k6x6x" event={"ID":"09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d","Type":"ContainerStarted","Data":"74747d63c37f6d97bd3c4866116fc0c78521a32c0c3d291c2c59f7ef80acb6a5"} Mar 20 09:52:02 crc kubenswrapper[4823]: I0320 09:52:02.529809 4823 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566672-k6x6x" podStartSLOduration=1.697667759 podStartE2EDuration="2.529737584s" podCreationTimestamp="2026-03-20 09:52:00 +0000 UTC" firstStartedPulling="2026-03-20 09:52:01.251134278 +0000 UTC m=+3287.937030692" lastFinishedPulling="2026-03-20 09:52:02.083204133 +0000 UTC m=+3288.769100517" observedRunningTime="2026-03-20 09:52:02.521577543 +0000 UTC m=+3289.207473917" watchObservedRunningTime="2026-03-20 09:52:02.529737584 +0000 UTC m=+3289.215634008" Mar 20 09:52:03 crc kubenswrapper[4823]: I0320 09:52:03.515802 4823 generic.go:334] "Generic (PLEG): container finished" podID="09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d" containerID="74747d63c37f6d97bd3c4866116fc0c78521a32c0c3d291c2c59f7ef80acb6a5" exitCode=0 Mar 20 09:52:03 crc kubenswrapper[4823]: I0320 09:52:03.515892 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566672-k6x6x" event={"ID":"09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d","Type":"ContainerDied","Data":"74747d63c37f6d97bd3c4866116fc0c78521a32c0c3d291c2c59f7ef80acb6a5"} Mar 20 09:52:04 crc kubenswrapper[4823]: I0320 09:52:04.778520 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566672-k6x6x" Mar 20 09:52:04 crc kubenswrapper[4823]: I0320 09:52:04.943985 4823 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9qbc\" (UniqueName: \"kubernetes.io/projected/09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d-kube-api-access-q9qbc\") pod \"09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d\" (UID: \"09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d\") " Mar 20 09:52:04 crc kubenswrapper[4823]: I0320 09:52:04.950261 4823 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d-kube-api-access-q9qbc" (OuterVolumeSpecName: "kube-api-access-q9qbc") pod "09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d" (UID: "09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d"). InnerVolumeSpecName "kube-api-access-q9qbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:52:05 crc kubenswrapper[4823]: I0320 09:52:05.046307 4823 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9qbc\" (UniqueName: \"kubernetes.io/projected/09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d-kube-api-access-q9qbc\") on node \"crc\" DevicePath \"\"" Mar 20 09:52:05 crc kubenswrapper[4823]: I0320 09:52:05.533175 4823 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566672-k6x6x" event={"ID":"09b45a43-7537-4ac3-bf6d-1f41fd2f5c1d","Type":"ContainerDied","Data":"57abb0b7f873fc862a148ca4409f89fca8f5a3d5c69c6e5527db83fc5c0b0674"} Mar 20 09:52:05 crc kubenswrapper[4823]: I0320 09:52:05.533448 4823 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57abb0b7f873fc862a148ca4409f89fca8f5a3d5c69c6e5527db83fc5c0b0674" Mar 20 09:52:05 crc kubenswrapper[4823]: I0320 09:52:05.533256 4823 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566672-k6x6x" Mar 20 09:52:05 crc kubenswrapper[4823]: I0320 09:52:05.592740 4823 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566666-np65s"] Mar 20 09:52:05 crc kubenswrapper[4823]: I0320 09:52:05.599533 4823 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566666-np65s"] Mar 20 09:52:06 crc kubenswrapper[4823]: I0320 09:52:06.105895 4823 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bb40728-21ed-496e-ac84-a39d9adaf100" path="/var/lib/kubelet/pods/0bb40728-21ed-496e-ac84-a39d9adaf100/volumes" Mar 20 09:52:40 crc kubenswrapper[4823]: I0320 09:52:40.791921 4823 scope.go:117] "RemoveContainer" containerID="66ee65d5905ec17f621581c57d0c5fddebd99633959722f32a839bf5ca7ded8c" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515157214444024454 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015157214445017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015157205450016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015157205450015461 5ustar corecore